var/home/core/zuul-output/0000755000175000017500000000000015070153352014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070171164015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005101406115070171156017676 0ustar rootrootOct 04 08:16:11 crc systemd[1]: Starting Kubernetes Kubelet... Oct 04 08:16:11 crc restorecon[4663]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:11 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 08:16:12 crc restorecon[4663]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 04 08:16:12 crc kubenswrapper[4969]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 08:16:12 crc kubenswrapper[4969]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 04 08:16:12 crc kubenswrapper[4969]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 08:16:12 crc kubenswrapper[4969]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 08:16:12 crc kubenswrapper[4969]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 04 08:16:12 crc kubenswrapper[4969]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.786075 4969 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792813 4969 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792845 4969 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792854 4969 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792863 4969 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792871 4969 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792879 4969 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792887 4969 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792895 4969 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792903 4969 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792910 4969 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792918 4969 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792926 4969 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792935 4969 feature_gate.go:330] unrecognized feature gate: Example Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792942 4969 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792950 4969 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792957 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792965 4969 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792973 4969 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792980 4969 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792988 4969 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.792995 4969 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793003 4969 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793010 4969 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793018 4969 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793026 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793033 4969 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793041 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793050 4969 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793058 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793066 4969 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793074 4969 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793082 4969 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793093 4969 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793104 4969 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793114 4969 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793122 4969 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793133 4969 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793143 4969 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793152 4969 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793161 4969 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793170 4969 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793179 4969 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793187 4969 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793196 4969 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793203 4969 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793232 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793241 4969 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793249 4969 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793258 4969 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793266 4969 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793276 4969 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793287 4969 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793295 4969 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793303 4969 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793313 4969 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793322 4969 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793330 4969 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793340 4969 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793348 4969 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793357 4969 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793366 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793374 4969 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793381 4969 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793389 4969 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793397 4969 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793404 4969 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793416 4969 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793452 4969 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793460 4969 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793468 4969 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.793475 4969 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793687 4969 flags.go:64] FLAG: --address="0.0.0.0" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793717 4969 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793733 4969 flags.go:64] FLAG: --anonymous-auth="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793744 4969 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793781 4969 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793791 4969 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793803 4969 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793815 4969 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793915 4969 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793926 4969 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793936 4969 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793946 4969 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793955 4969 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793964 4969 flags.go:64] FLAG: --cgroup-root="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793973 4969 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793982 4969 flags.go:64] FLAG: --client-ca-file="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.793991 4969 flags.go:64] FLAG: --cloud-config="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794000 4969 flags.go:64] FLAG: --cloud-provider="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794019 4969 flags.go:64] FLAG: --cluster-dns="[]" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794032 4969 flags.go:64] FLAG: --cluster-domain="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794041 4969 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794050 4969 flags.go:64] FLAG: --config-dir="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794059 4969 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794068 4969 flags.go:64] FLAG: --container-log-max-files="5" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794089 4969 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794099 4969 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794108 4969 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794117 4969 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794126 4969 flags.go:64] FLAG: --contention-profiling="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794135 4969 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794145 4969 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794154 4969 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794164 4969 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794175 4969 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794184 4969 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794194 4969 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794203 4969 flags.go:64] FLAG: --enable-load-reader="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794212 4969 flags.go:64] FLAG: --enable-server="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794221 4969 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794232 4969 flags.go:64] FLAG: --event-burst="100" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794241 4969 flags.go:64] FLAG: --event-qps="50" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794251 4969 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794260 4969 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794269 4969 flags.go:64] FLAG: --eviction-hard="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794280 4969 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794289 4969 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794298 4969 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794308 4969 flags.go:64] FLAG: --eviction-soft="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794317 4969 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794325 4969 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794335 4969 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794344 4969 flags.go:64] FLAG: --experimental-mounter-path="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794353 4969 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794361 4969 flags.go:64] FLAG: --fail-swap-on="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794370 4969 flags.go:64] FLAG: --feature-gates="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794381 4969 flags.go:64] FLAG: --file-check-frequency="20s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794390 4969 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794399 4969 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794409 4969 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794451 4969 flags.go:64] FLAG: --healthz-port="10248" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794461 4969 flags.go:64] FLAG: --help="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794470 4969 flags.go:64] FLAG: --hostname-override="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794479 4969 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794488 4969 flags.go:64] FLAG: --http-check-frequency="20s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794498 4969 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794506 4969 flags.go:64] FLAG: --image-credential-provider-config="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794516 4969 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794524 4969 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794533 4969 flags.go:64] FLAG: --image-service-endpoint="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794542 4969 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794551 4969 flags.go:64] FLAG: --kube-api-burst="100" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794560 4969 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794570 4969 flags.go:64] FLAG: --kube-api-qps="50" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794578 4969 flags.go:64] FLAG: --kube-reserved="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794587 4969 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794596 4969 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794605 4969 flags.go:64] FLAG: --kubelet-cgroups="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794614 4969 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794623 4969 flags.go:64] FLAG: --lock-file="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794632 4969 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794642 4969 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794651 4969 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794665 4969 flags.go:64] FLAG: --log-json-split-stream="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794674 4969 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794683 4969 flags.go:64] FLAG: --log-text-split-stream="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794692 4969 flags.go:64] FLAG: --logging-format="text" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794701 4969 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794711 4969 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794720 4969 flags.go:64] FLAG: --manifest-url="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794729 4969 flags.go:64] FLAG: --manifest-url-header="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794740 4969 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794749 4969 flags.go:64] FLAG: --max-open-files="1000000" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794761 4969 flags.go:64] FLAG: --max-pods="110" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794770 4969 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794779 4969 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794787 4969 flags.go:64] FLAG: --memory-manager-policy="None" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794796 4969 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794806 4969 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794815 4969 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794824 4969 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794845 4969 flags.go:64] FLAG: --node-status-max-images="50" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794854 4969 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794863 4969 flags.go:64] FLAG: --oom-score-adj="-999" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794872 4969 flags.go:64] FLAG: --pod-cidr="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794881 4969 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794892 4969 flags.go:64] FLAG: --pod-manifest-path="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794901 4969 flags.go:64] FLAG: --pod-max-pids="-1" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794910 4969 flags.go:64] FLAG: --pods-per-core="0" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794919 4969 flags.go:64] FLAG: --port="10250" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794929 4969 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794937 4969 flags.go:64] FLAG: --provider-id="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794946 4969 flags.go:64] FLAG: --qos-reserved="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794955 4969 flags.go:64] FLAG: --read-only-port="10255" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794964 4969 flags.go:64] FLAG: --register-node="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794973 4969 flags.go:64] FLAG: --register-schedulable="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794981 4969 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.794996 4969 flags.go:64] FLAG: --registry-burst="10" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795006 4969 flags.go:64] FLAG: --registry-qps="5" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795015 4969 flags.go:64] FLAG: --reserved-cpus="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795023 4969 flags.go:64] FLAG: --reserved-memory="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795034 4969 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795043 4969 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795052 4969 flags.go:64] FLAG: --rotate-certificates="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795061 4969 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795070 4969 flags.go:64] FLAG: --runonce="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795079 4969 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795088 4969 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795098 4969 flags.go:64] FLAG: --seccomp-default="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795107 4969 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795116 4969 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795126 4969 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795135 4969 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795144 4969 flags.go:64] FLAG: --storage-driver-password="root" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795154 4969 flags.go:64] FLAG: --storage-driver-secure="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795164 4969 flags.go:64] FLAG: --storage-driver-table="stats" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795172 4969 flags.go:64] FLAG: --storage-driver-user="root" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795181 4969 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795190 4969 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795200 4969 flags.go:64] FLAG: --system-cgroups="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795209 4969 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795223 4969 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795232 4969 flags.go:64] FLAG: --tls-cert-file="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795240 4969 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795252 4969 flags.go:64] FLAG: --tls-min-version="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795261 4969 flags.go:64] FLAG: --tls-private-key-file="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795271 4969 flags.go:64] FLAG: --topology-manager-policy="none" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795280 4969 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795288 4969 flags.go:64] FLAG: --topology-manager-scope="container" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795297 4969 flags.go:64] FLAG: --v="2" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795318 4969 flags.go:64] FLAG: --version="false" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795333 4969 flags.go:64] FLAG: --vmodule="" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795344 4969 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.795354 4969 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795589 4969 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795601 4969 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795610 4969 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795618 4969 feature_gate.go:330] unrecognized feature gate: Example Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795626 4969 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795634 4969 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795643 4969 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795650 4969 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795659 4969 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795667 4969 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795675 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795683 4969 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795691 4969 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795699 4969 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795710 4969 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795719 4969 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795727 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795736 4969 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795745 4969 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795754 4969 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795762 4969 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795770 4969 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795777 4969 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795785 4969 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795793 4969 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795801 4969 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795809 4969 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795817 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795825 4969 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795835 4969 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795843 4969 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795853 4969 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795865 4969 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795875 4969 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795884 4969 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795894 4969 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795903 4969 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795910 4969 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795919 4969 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795926 4969 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795934 4969 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795942 4969 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795950 4969 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795958 4969 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795966 4969 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795974 4969 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795982 4969 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795989 4969 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.795997 4969 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796005 4969 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796013 4969 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796021 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796028 4969 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796036 4969 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796047 4969 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796056 4969 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796065 4969 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796073 4969 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796087 4969 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796095 4969 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796103 4969 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796113 4969 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796121 4969 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796129 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796137 4969 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796145 4969 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796152 4969 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796160 4969 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796168 4969 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796175 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.796183 4969 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.796207 4969 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.808404 4969 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.808524 4969 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808681 4969 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808717 4969 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808728 4969 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808737 4969 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808746 4969 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808757 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808768 4969 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808777 4969 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808787 4969 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808797 4969 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808806 4969 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808816 4969 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808826 4969 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808836 4969 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808846 4969 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808856 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808866 4969 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808879 4969 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808894 4969 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808905 4969 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808916 4969 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808927 4969 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808938 4969 feature_gate.go:330] unrecognized feature gate: Example Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808948 4969 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808958 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808968 4969 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808978 4969 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.808989 4969 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809003 4969 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809016 4969 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809028 4969 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809039 4969 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809050 4969 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809064 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809075 4969 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809086 4969 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809096 4969 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809108 4969 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809119 4969 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809130 4969 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809140 4969 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809151 4969 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809161 4969 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809171 4969 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809182 4969 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809191 4969 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809201 4969 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809210 4969 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809255 4969 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809268 4969 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809279 4969 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809290 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809299 4969 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809309 4969 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809318 4969 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809327 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809335 4969 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809343 4969 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809352 4969 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809359 4969 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809367 4969 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809375 4969 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809382 4969 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809390 4969 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809400 4969 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809410 4969 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809449 4969 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809458 4969 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809466 4969 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809477 4969 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809485 4969 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.809498 4969 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809734 4969 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809746 4969 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809755 4969 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809763 4969 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809772 4969 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809779 4969 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809787 4969 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809795 4969 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809802 4969 feature_gate.go:330] unrecognized feature gate: Example Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809810 4969 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809818 4969 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809826 4969 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809834 4969 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809841 4969 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809849 4969 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809858 4969 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809867 4969 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809878 4969 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809887 4969 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809898 4969 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809908 4969 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809918 4969 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809929 4969 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809938 4969 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809947 4969 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809957 4969 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809966 4969 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809977 4969 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809986 4969 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.809996 4969 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810006 4969 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810015 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810029 4969 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810045 4969 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810057 4969 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810067 4969 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810077 4969 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810087 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810095 4969 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810103 4969 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810110 4969 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810118 4969 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810125 4969 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810134 4969 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810141 4969 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810150 4969 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810161 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810177 4969 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810228 4969 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810241 4969 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810251 4969 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810261 4969 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810270 4969 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810278 4969 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810287 4969 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810295 4969 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810304 4969 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810312 4969 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810319 4969 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810328 4969 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810336 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810346 4969 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810355 4969 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810363 4969 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810371 4969 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810380 4969 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810387 4969 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810395 4969 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810405 4969 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810447 4969 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.810456 4969 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.810469 4969 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.811722 4969 server.go:940] "Client rotation is on, will bootstrap in background" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.817593 4969 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.817742 4969 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.819975 4969 server.go:997] "Starting client certificate rotation" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.820026 4969 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.820969 4969 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-10 06:09:07.920011365 +0000 UTC Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.821059 4969 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 885h52m55.098954624s for next certificate rotation Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.848363 4969 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.851288 4969 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.868205 4969 log.go:25] "Validated CRI v1 runtime API" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.911510 4969 log.go:25] "Validated CRI v1 image API" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.915961 4969 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.921457 4969 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-04-08-11-17-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.921506 4969 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.948838 4969 manager.go:217] Machine: {Timestamp:2025-10-04 08:16:12.945526722 +0000 UTC m=+0.699795616 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:840d234f-97b2-4b81-bc5d-e0ca9b13faeb BootID:529892e6-c212-481c-813c-a422ee3e502d Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ee:8f:ca Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ee:8f:ca Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:de:43:0b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:6b:97:c8 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:3c:41:8a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:47:87:84 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:3e:20:0f:d4:ef:78 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:66:07:f4:29:e9:63 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.949518 4969 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.949851 4969 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.951491 4969 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.952036 4969 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.952258 4969 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.952861 4969 topology_manager.go:138] "Creating topology manager with none policy" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.953025 4969 container_manager_linux.go:303] "Creating device plugin manager" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.953686 4969 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.953921 4969 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.954592 4969 state_mem.go:36] "Initialized new in-memory state store" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.954952 4969 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.958606 4969 kubelet.go:418] "Attempting to sync node with API server" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.958766 4969 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.959010 4969 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.959161 4969 kubelet.go:324] "Adding apiserver pod source" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.959283 4969 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.963994 4969 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.965144 4969 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.966227 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:12 crc kubenswrapper[4969]: E1004 08:16:12.966564 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.966366 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:12 crc kubenswrapper[4969]: E1004 08:16:12.966963 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.967118 4969 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968623 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968663 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968678 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968690 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968712 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968725 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968738 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968771 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968799 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968819 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968890 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.968906 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.969913 4969 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.970556 4969 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.970636 4969 server.go:1280] "Started kubelet" Oct 04 08:16:12 crc systemd[1]: Started Kubernetes Kubelet. Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.971750 4969 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.971750 4969 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.979733 4969 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.982337 4969 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.982453 4969 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.982581 4969 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.982616 4969 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.982653 4969 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 19:18:24.042254187 +0000 UTC Oct 04 08:16:12 crc kubenswrapper[4969]: E1004 08:16:12.982797 4969 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.983017 4969 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1547h2m11.059496216s for next certificate rotation Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.983186 4969 server.go:460] "Adding debug handlers to kubelet server" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.983297 4969 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 04 08:16:12 crc kubenswrapper[4969]: W1004 08:16:12.984504 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:12 crc kubenswrapper[4969]: E1004 08:16:12.984626 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.984741 4969 factory.go:55] Registering systemd factory Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.984787 4969 factory.go:221] Registration of the systemd container factory successfully Oct 04 08:16:12 crc kubenswrapper[4969]: E1004 08:16:12.985403 4969 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="200ms" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.988303 4969 factory.go:153] Registering CRI-O factory Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.988806 4969 factory.go:221] Registration of the crio container factory successfully Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.988943 4969 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.988987 4969 factory.go:103] Registering Raw factory Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.989025 4969 manager.go:1196] Started watching for new ooms in manager Oct 04 08:16:12 crc kubenswrapper[4969]: E1004 08:16:12.987886 4969 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.195:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b3bbf78d92aed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 08:16:12.970601197 +0000 UTC m=+0.724870041,LastTimestamp:2025-10-04 08:16:12.970601197 +0000 UTC m=+0.724870041,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 08:16:12 crc kubenswrapper[4969]: I1004 08:16:12.990175 4969 manager.go:319] Starting recovery of all containers Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000379 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000494 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000519 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000539 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000557 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000578 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000596 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000616 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000679 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000698 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000719 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000744 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000769 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000802 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000854 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000883 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000902 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000926 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000943 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000961 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.000978 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001002 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001020 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001037 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001184 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001205 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001270 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001294 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001312 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001329 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001382 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001401 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001444 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001467 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001487 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001506 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001523 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001542 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001558 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001575 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001593 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001611 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001630 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001649 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001668 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001685 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001702 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001720 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001743 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001768 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001797 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001818 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001843 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001864 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001888 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001912 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001935 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001958 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.001983 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002005 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002028 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002064 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002086 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002110 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002133 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002154 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002175 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002200 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002225 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002248 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002271 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002296 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002320 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002342 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002364 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002388 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002414 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002475 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002499 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002526 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002550 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002577 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002601 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002626 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002650 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002688 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002713 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002738 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002775 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002800 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002824 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002940 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002966 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.002990 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.003011 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.003036 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.003060 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.003087 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.003110 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.003133 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.003158 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007187 4969 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007263 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007296 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007329 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007373 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007405 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007469 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007500 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007564 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007595 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007629 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007657 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007691 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007720 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007749 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007803 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007830 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007858 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007890 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007918 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007946 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.007970 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008000 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008088 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008114 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008140 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008166 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008190 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008217 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008248 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008273 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008297 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008324 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008347 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008370 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008393 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008452 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008483 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008509 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008534 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008557 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008583 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008608 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008634 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008663 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008690 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008715 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008754 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008780 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008807 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008834 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008858 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008886 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008913 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008942 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008966 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.008991 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009016 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009041 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009066 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009096 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009122 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009147 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009174 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009199 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009266 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009293 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009320 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009345 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009370 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009395 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009462 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009492 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009516 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009542 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009568 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009599 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009625 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009650 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009677 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009704 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009731 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009759 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009785 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009825 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009859 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009888 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009914 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009937 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009964 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.009990 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010015 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010041 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010069 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010093 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010118 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010143 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010165 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010192 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010215 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010238 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010259 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010281 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010307 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010332 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010358 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010383 4969 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010410 4969 reconstruct.go:97] "Volume reconstruction finished" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.010469 4969 reconciler.go:26] "Reconciler: start to sync state" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.024649 4969 manager.go:324] Recovery completed Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.043742 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.045758 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.045798 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.045809 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.047156 4969 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.047256 4969 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.047339 4969 state_mem.go:36] "Initialized new in-memory state store" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.050798 4969 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.053809 4969 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.053884 4969 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.053918 4969 kubelet.go:2335] "Starting kubelet main sync loop" Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.054097 4969 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 04 08:16:13 crc kubenswrapper[4969]: W1004 08:16:13.054628 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.054757 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.064920 4969 policy_none.go:49] "None policy: Start" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.066092 4969 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.066129 4969 state_mem.go:35] "Initializing new in-memory state store" Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.083932 4969 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.133134 4969 manager.go:334] "Starting Device Plugin manager" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.133331 4969 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.133348 4969 server.go:79] "Starting device plugin registration server" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.133788 4969 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.133806 4969 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.134010 4969 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.134193 4969 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.134210 4969 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.143213 4969 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.154917 4969 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.155007 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.156676 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.156714 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.156725 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.156845 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.157011 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.157041 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.157691 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.157723 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.157733 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.157828 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.157975 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.158008 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.158398 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.158466 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.158479 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.158912 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.158939 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.158949 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.158909 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.159000 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.159023 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.159052 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.159145 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.159173 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.159780 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.159798 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.159809 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.159897 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160064 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160107 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160110 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160144 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160152 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160554 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160581 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160591 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160719 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160743 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160920 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.160999 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.161025 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.161391 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.161456 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.161470 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.186906 4969 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="400ms" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213331 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213377 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213403 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213451 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213483 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213508 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213579 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213653 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213711 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213747 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213778 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213812 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213860 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213908 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.213933 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.234399 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.235935 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.235979 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.235991 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.236018 4969 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.236556 4969 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315506 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315534 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315707 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315808 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315881 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315804 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315914 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315923 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315978 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.315956 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316055 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316071 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316016 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316144 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316178 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316207 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316212 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316179 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316245 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316247 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316271 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316252 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316310 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316332 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316341 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316358 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316372 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316401 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316441 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.316522 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.436694 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.438294 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.438385 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.438407 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.438480 4969 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.438974 4969 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.500311 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.522166 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.532400 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.554065 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: W1004 08:16:13.555337 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-7ed244110e3277c65c2827ad0311c0abfb54e4aba0e159bbe01baad931c74777 WatchSource:0}: Error finding container 7ed244110e3277c65c2827ad0311c0abfb54e4aba0e159bbe01baad931c74777: Status 404 returned error can't find the container with id 7ed244110e3277c65c2827ad0311c0abfb54e4aba0e159bbe01baad931c74777 Oct 04 08:16:13 crc kubenswrapper[4969]: W1004 08:16:13.562195 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5a37e021b9d911744a367b9e9e6b35083128c838d124c02a3ef24db085116bd1 WatchSource:0}: Error finding container 5a37e021b9d911744a367b9e9e6b35083128c838d124c02a3ef24db085116bd1: Status 404 returned error can't find the container with id 5a37e021b9d911744a367b9e9e6b35083128c838d124c02a3ef24db085116bd1 Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.564829 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:16:13 crc kubenswrapper[4969]: W1004 08:16:13.566629 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-88bc1c9cb7815b4d16dde5ab2577d1f3b6ba4c945f6d151b92afb85b62af1c16 WatchSource:0}: Error finding container 88bc1c9cb7815b4d16dde5ab2577d1f3b6ba4c945f6d151b92afb85b62af1c16: Status 404 returned error can't find the container with id 88bc1c9cb7815b4d16dde5ab2577d1f3b6ba4c945f6d151b92afb85b62af1c16 Oct 04 08:16:13 crc kubenswrapper[4969]: W1004 08:16:13.575867 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-d6c55bcea2201570842ad8973d2eb48acc73a1e545e05bc1ab1e635c6e4e0892 WatchSource:0}: Error finding container d6c55bcea2201570842ad8973d2eb48acc73a1e545e05bc1ab1e635c6e4e0892: Status 404 returned error can't find the container with id d6c55bcea2201570842ad8973d2eb48acc73a1e545e05bc1ab1e635c6e4e0892 Oct 04 08:16:13 crc kubenswrapper[4969]: W1004 08:16:13.586912 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-fb9d845fcee68e4c98c249b3a2c3bd051197a538dca09ab737754048004a8820 WatchSource:0}: Error finding container fb9d845fcee68e4c98c249b3a2c3bd051197a538dca09ab737754048004a8820: Status 404 returned error can't find the container with id fb9d845fcee68e4c98c249b3a2c3bd051197a538dca09ab737754048004a8820 Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.587737 4969 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="800ms" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.839257 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.840935 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.840988 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.841007 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.841042 4969 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.841697 4969 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Oct 04 08:16:13 crc kubenswrapper[4969]: W1004 08:16:13.907137 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.907251 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:13 crc kubenswrapper[4969]: W1004 08:16:13.925348 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:13 crc kubenswrapper[4969]: E1004 08:16:13.925497 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:13 crc kubenswrapper[4969]: I1004 08:16:13.972085 4969 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.063096 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d6c55bcea2201570842ad8973d2eb48acc73a1e545e05bc1ab1e635c6e4e0892"} Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.065637 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"88bc1c9cb7815b4d16dde5ab2577d1f3b6ba4c945f6d151b92afb85b62af1c16"} Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.066670 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5a37e021b9d911744a367b9e9e6b35083128c838d124c02a3ef24db085116bd1"} Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.069275 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7ed244110e3277c65c2827ad0311c0abfb54e4aba0e159bbe01baad931c74777"} Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.070383 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fb9d845fcee68e4c98c249b3a2c3bd051197a538dca09ab737754048004a8820"} Oct 04 08:16:14 crc kubenswrapper[4969]: W1004 08:16:14.372613 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:14 crc kubenswrapper[4969]: E1004 08:16:14.373006 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:14 crc kubenswrapper[4969]: E1004 08:16:14.389266 4969 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="1.6s" Oct 04 08:16:14 crc kubenswrapper[4969]: W1004 08:16:14.480997 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:14 crc kubenswrapper[4969]: E1004 08:16:14.481172 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.642241 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.643955 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.644005 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.644028 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.644061 4969 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 08:16:14 crc kubenswrapper[4969]: E1004 08:16:14.644772 4969 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Oct 04 08:16:14 crc kubenswrapper[4969]: I1004 08:16:14.971733 4969 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.075648 4969 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff" exitCode=0 Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.075785 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff"} Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.075797 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.077758 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.077810 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.077828 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.079888 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.081058 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.081121 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.081138 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.081872 4969 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5b8c5a7323a8b918389703f6525573c8f745dde7e6b42014cb4696d8de887038" exitCode=0 Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.081953 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.081967 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5b8c5a7323a8b918389703f6525573c8f745dde7e6b42014cb4696d8de887038"} Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.083038 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.083065 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.083082 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.084736 4969 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a" exitCode=0 Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.084823 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a"} Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.084859 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.087150 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.087239 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.087329 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.088037 4969 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b" exitCode=0 Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.088123 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b"} Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.088194 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.089938 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.089997 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.090020 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.093049 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b"} Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.093086 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5"} Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.093100 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7"} Oct 04 08:16:15 crc kubenswrapper[4969]: I1004 08:16:15.971959 4969 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:15 crc kubenswrapper[4969]: E1004 08:16:15.990234 4969 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="3.2s" Oct 04 08:16:16 crc kubenswrapper[4969]: W1004 08:16:16.069792 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:16 crc kubenswrapper[4969]: E1004 08:16:16.069868 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.099300 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.099337 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.100383 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.100460 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.100474 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.107633 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.107691 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.107712 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.107728 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.117272 4969 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="18e41bf563ad1f3e24583842ac5c3931e1fcbfeecb5f3ca368eab9043b2630f4" exitCode=0 Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.117387 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"18e41bf563ad1f3e24583842ac5c3931e1fcbfeecb5f3ca368eab9043b2630f4"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.117438 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.118322 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.118355 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.118366 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.121225 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8a9c80c537af220b0de87725a1be7e5d7370abd4639e79d49afd8226718aaeed"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.121268 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.122180 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.122202 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.122212 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.124207 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.124252 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.124272 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27"} Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.124357 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.125260 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.125312 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.125324 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.245922 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.247117 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.247159 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.247168 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:16 crc kubenswrapper[4969]: I1004 08:16:16.247192 4969 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 08:16:16 crc kubenswrapper[4969]: E1004 08:16:16.247635 4969 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Oct 04 08:16:16 crc kubenswrapper[4969]: W1004 08:16:16.419681 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Oct 04 08:16:16 crc kubenswrapper[4969]: E1004 08:16:16.419751 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.132630 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183"} Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.132720 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.134342 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.134447 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.134468 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.136722 4969 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7d374408c0b4c963b5aa28c2dd232863da4c8da5d115b004ea3a1b6b60ea61ac" exitCode=0 Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.136832 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.136845 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7d374408c0b4c963b5aa28c2dd232863da4c8da5d115b004ea3a1b6b60ea61ac"} Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.136870 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.136974 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.136988 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.136874 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139021 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139060 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139101 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139121 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139073 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139199 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139198 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139251 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139071 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139307 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139333 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:17 crc kubenswrapper[4969]: I1004 08:16:17.139276 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.145730 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1fb87050241de09cb0d7fcc9cf92e5a312461fa3e04a3eedf952e5dbcf28f770"} Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.145792 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"93d8ad417607834b494bbee17c0ce49dcf82732b7b62e9758788172146763876"} Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.145808 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.145921 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.145812 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d9b14d2dda186e85ea5064452bfe21733b6e320c7485b88d22bad57bad45e030"} Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.147397 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.147488 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.147673 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.280535 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.280817 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.282612 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.282673 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.282692 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.561891 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.562226 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.564353 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.564455 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.564473 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:18 crc kubenswrapper[4969]: I1004 08:16:18.571401 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.157754 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"512c1c55de5dd59495dbb29929165c1806d5882e971f44141d3db93b5660ba61"} Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.157861 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4fbca2af7000d931ff48567f17a5426141de93c637621b69057c570a15d18653"} Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.158044 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.159617 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.159694 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.159719 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.160124 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.163211 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.163269 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.163286 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.447848 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.450186 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.450257 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.450282 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.450324 4969 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 08:16:19 crc kubenswrapper[4969]: I1004 08:16:19.902343 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.160023 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.160108 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.161513 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.161568 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.161591 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.161840 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.161899 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.161917 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.531609 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.531850 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.531907 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.533724 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.533786 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.533805 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.575490 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:20 crc kubenswrapper[4969]: I1004 08:16:20.745525 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:21 crc kubenswrapper[4969]: I1004 08:16:21.162360 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:21 crc kubenswrapper[4969]: I1004 08:16:21.164138 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:21 crc kubenswrapper[4969]: I1004 08:16:21.164203 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:21 crc kubenswrapper[4969]: I1004 08:16:21.164223 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:21 crc kubenswrapper[4969]: I1004 08:16:21.952901 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 04 08:16:21 crc kubenswrapper[4969]: I1004 08:16:21.953162 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:21 crc kubenswrapper[4969]: I1004 08:16:21.954679 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:21 crc kubenswrapper[4969]: I1004 08:16:21.954731 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:21 crc kubenswrapper[4969]: I1004 08:16:21.954747 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:22 crc kubenswrapper[4969]: I1004 08:16:22.165686 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:22 crc kubenswrapper[4969]: I1004 08:16:22.167010 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:22 crc kubenswrapper[4969]: I1004 08:16:22.167089 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:22 crc kubenswrapper[4969]: I1004 08:16:22.167108 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:22 crc kubenswrapper[4969]: I1004 08:16:22.221071 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:22 crc kubenswrapper[4969]: I1004 08:16:22.221293 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:22 crc kubenswrapper[4969]: I1004 08:16:22.222677 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:22 crc kubenswrapper[4969]: I1004 08:16:22.222766 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:22 crc kubenswrapper[4969]: I1004 08:16:22.222793 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:23 crc kubenswrapper[4969]: E1004 08:16:23.143414 4969 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.089410 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.089753 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.091309 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.091375 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.091397 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.097893 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.177408 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.179305 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.179400 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.179458 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:26 crc kubenswrapper[4969]: W1004 08:16:26.731037 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.731207 4969 trace.go:236] Trace[541790944]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 08:16:16.729) (total time: 10001ms): Oct 04 08:16:26 crc kubenswrapper[4969]: Trace[541790944]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:16:26.731) Oct 04 08:16:26 crc kubenswrapper[4969]: Trace[541790944]: [10.001602305s] [10.001602305s] END Oct 04 08:16:26 crc kubenswrapper[4969]: E1004 08:16:26.731255 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 04 08:16:26 crc kubenswrapper[4969]: W1004 08:16:26.786293 4969 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.786580 4969 trace.go:236] Trace[1529448200]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 08:16:16.785) (total time: 10001ms): Oct 04 08:16:26 crc kubenswrapper[4969]: Trace[1529448200]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:16:26.786) Oct 04 08:16:26 crc kubenswrapper[4969]: Trace[1529448200]: [10.001509391s] [10.001509391s] END Oct 04 08:16:26 crc kubenswrapper[4969]: E1004 08:16:26.786835 4969 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 04 08:16:26 crc kubenswrapper[4969]: I1004 08:16:26.972266 4969 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.295361 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.295652 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.296922 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.296981 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.297003 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.353204 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.513533 4969 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.513655 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.530155 4969 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 08:16:27 crc kubenswrapper[4969]: I1004 08:16:27.530232 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 08:16:28 crc kubenswrapper[4969]: I1004 08:16:28.182010 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:28 crc kubenswrapper[4969]: I1004 08:16:28.182973 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:28 crc kubenswrapper[4969]: I1004 08:16:28.183035 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:28 crc kubenswrapper[4969]: I1004 08:16:28.183055 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:28 crc kubenswrapper[4969]: I1004 08:16:28.205986 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 04 08:16:29 crc kubenswrapper[4969]: I1004 08:16:29.089416 4969 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 08:16:29 crc kubenswrapper[4969]: I1004 08:16:29.089583 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 08:16:29 crc kubenswrapper[4969]: I1004 08:16:29.184779 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:29 crc kubenswrapper[4969]: I1004 08:16:29.185865 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:29 crc kubenswrapper[4969]: I1004 08:16:29.185916 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:29 crc kubenswrapper[4969]: I1004 08:16:29.185935 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:30 crc kubenswrapper[4969]: I1004 08:16:30.753324 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:30 crc kubenswrapper[4969]: I1004 08:16:30.753546 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:30 crc kubenswrapper[4969]: I1004 08:16:30.755134 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:30 crc kubenswrapper[4969]: I1004 08:16:30.755236 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:30 crc kubenswrapper[4969]: I1004 08:16:30.755254 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:30 crc kubenswrapper[4969]: I1004 08:16:30.757957 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:31 crc kubenswrapper[4969]: I1004 08:16:31.190196 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:16:31 crc kubenswrapper[4969]: I1004 08:16:31.190650 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:31 crc kubenswrapper[4969]: I1004 08:16:31.192261 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:31 crc kubenswrapper[4969]: I1004 08:16:31.192325 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:31 crc kubenswrapper[4969]: I1004 08:16:31.192342 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.065167 4969 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.094994 4969 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.509387 4969 trace.go:236] Trace[1850604115]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 08:16:22.121) (total time: 10387ms): Oct 04 08:16:32 crc kubenswrapper[4969]: Trace[1850604115]: ---"Objects listed" error: 10387ms (08:16:32.509) Oct 04 08:16:32 crc kubenswrapper[4969]: Trace[1850604115]: [10.387469721s] [10.387469721s] END Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.509445 4969 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 04 08:16:32 crc kubenswrapper[4969]: E1004 08:16:32.512878 4969 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.516231 4969 trace.go:236] Trace[578904516]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 08:16:21.478) (total time: 11037ms): Oct 04 08:16:32 crc kubenswrapper[4969]: Trace[578904516]: ---"Objects listed" error: 11037ms (08:16:32.516) Oct 04 08:16:32 crc kubenswrapper[4969]: Trace[578904516]: [11.037311464s] [11.037311464s] END Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.516259 4969 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.517747 4969 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 04 08:16:32 crc kubenswrapper[4969]: E1004 08:16:32.518894 4969 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.572056 4969 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60642->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.572098 4969 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60628->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.572154 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60628->192.168.126.11:17697: read: connection reset by peer" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.572106 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60642->192.168.126.11:17697: read: connection reset by peer" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.572533 4969 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.572564 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.971057 4969 apiserver.go:52] "Watching apiserver" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.973817 4969 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.974065 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.974531 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.974589 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.974646 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.974552 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.974692 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.974825 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 08:16:32 crc kubenswrapper[4969]: E1004 08:16:32.974953 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:32 crc kubenswrapper[4969]: E1004 08:16:32.975050 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:32 crc kubenswrapper[4969]: E1004 08:16:32.975126 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.976399 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.977409 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.977596 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.977724 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.977789 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.978068 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.978163 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.978302 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.978315 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 04 08:16:32 crc kubenswrapper[4969]: I1004 08:16:32.983917 4969 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.014902 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.020955 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021020 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021053 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021084 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021113 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021142 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021191 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021227 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021261 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021291 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021320 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021348 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021376 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021404 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021458 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021490 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021520 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021526 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021535 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021555 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021628 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021662 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021687 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021710 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021763 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021778 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021794 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021811 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021826 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021841 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021858 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021874 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021890 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021904 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021920 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021936 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021953 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021977 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021999 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022016 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022034 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022055 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022074 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022115 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022160 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022183 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022204 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022250 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022279 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022298 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022339 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022358 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022392 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022436 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022667 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022696 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022718 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022739 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022764 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022782 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022799 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022884 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022935 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022953 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022969 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021591 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.024696 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:16:33.524384475 +0000 UTC m=+21.278653289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.024753 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.024781 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.024967 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.024986 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021762 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021959 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025053 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.021970 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022146 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022347 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.022334 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.023046 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.023172 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.023336 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.023385 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.023409 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.023524 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.023739 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.023809 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.023869 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.024025 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.024091 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.024291 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025117 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025263 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025291 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025468 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025480 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.024378 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025512 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025545 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025576 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025608 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025628 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025648 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025669 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025690 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025712 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025735 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025757 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025780 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025806 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025833 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025853 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025874 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025894 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025922 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025943 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025961 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.025983 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026006 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026029 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026052 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026073 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026073 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026094 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026118 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026141 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026158 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026173 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026188 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026210 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026226 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026241 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026261 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026258 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026277 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026363 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026411 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026475 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026507 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026546 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026580 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026613 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026646 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026678 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026680 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026725 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026745 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026763 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026780 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026798 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026816 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026834 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026850 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026866 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026883 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.026901 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027013 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027030 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027046 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027062 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027080 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027095 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027111 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027128 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027145 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027118 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027160 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027147 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027196 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027300 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027518 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027161 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027580 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027605 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027603 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027645 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027668 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027724 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027743 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027760 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027756 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027778 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027796 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027840 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027863 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027887 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027909 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027927 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027945 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027963 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027978 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.027993 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028010 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028028 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028046 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028063 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028077 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028094 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028112 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028127 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028142 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028158 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028176 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028191 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028207 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028224 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028246 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028271 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028292 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028315 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028333 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028353 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028370 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028385 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028405 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028446 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028468 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028488 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028504 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028520 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028535 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028555 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028599 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028614 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028636 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028653 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028670 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028686 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028702 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028718 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028736 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028731 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028754 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028815 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028866 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028915 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028911 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028954 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.028980 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029005 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029034 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029062 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029086 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029141 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029172 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029203 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029234 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029246 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029269 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029301 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029329 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029355 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029381 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029406 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029455 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029482 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029486 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029515 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029542 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029608 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029615 4969 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029637 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029643 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029658 4969 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029673 4969 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029687 4969 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029701 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029702 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029715 4969 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029728 4969 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029743 4969 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029756 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029770 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029784 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029797 4969 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029811 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029824 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029837 4969 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029838 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029851 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029866 4969 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029868 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029881 4969 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029894 4969 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029908 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029921 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029937 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029950 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029965 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029980 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.029994 4969 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030006 4969 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030550 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030567 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030584 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030599 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030612 4969 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030626 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030641 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030659 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030672 4969 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030687 4969 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030700 4969 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030714 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030728 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030745 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030766 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030785 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030805 4969 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030820 4969 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030008 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030849 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030033 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030073 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030227 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030257 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030300 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030323 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.030360 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.031556 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.031582 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.031987 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.032184 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.032331 4969 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.032485 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.032794 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.032944 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.033093 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.033157 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.033217 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.033377 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.033467 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.033610 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.033729 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.034587 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.035028 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.035138 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.035319 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.038439 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.038708 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.038739 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.038890 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.039196 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.039503 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.039766 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.040125 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.040668 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.040715 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.040938 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.041132 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.041338 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.041855 4969 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.041858 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.041938 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:33.541916437 +0000 UTC m=+21.296185351 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.042090 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.042358 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.042528 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.042623 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.042639 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.042651 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.042673 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.042969 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.043092 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.043363 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.043474 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.043502 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.043608 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.043816 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.044113 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.044291 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.044473 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.044559 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.044586 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.041176 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.044809 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.045647 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.045722 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.046007 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.046501 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.046655 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.046850 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.046890 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.047202 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.047271 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.047451 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.047839 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.048094 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.048591 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.048850 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.048857 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.048836 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.049027 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.050080 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.050286 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.050468 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.050730 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.050836 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.051139 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.051195 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.051205 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.051355 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.051407 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.051779 4969 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.051781 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.052047 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.052067 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.052085 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.052102 4969 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.052158 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:33.552135863 +0000 UTC m=+21.306404787 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.052453 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:33.55243987 +0000 UTC m=+21.306708784 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.052817 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.052871 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.053011 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.053308 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.053480 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.053762 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.054106 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.054160 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.054629 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.054719 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.054747 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.056206 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.056577 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.058199 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.059827 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.061036 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.061378 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.061520 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.061400 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.061564 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.061605 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.061622 4969 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.061619 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.061676 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:33.561656242 +0000 UTC m=+21.315925056 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.062620 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.063728 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.064613 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.064973 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.065815 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.066804 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.067116 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.067584 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.068620 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.069028 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.069366 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.069517 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.070235 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.070296 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.070342 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.070578 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.070601 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.070832 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.070978 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.071387 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.071173 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.071263 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.071337 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.072823 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.073080 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.073389 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.073607 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.073651 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.073803 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.073961 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.074743 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.074966 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.075243 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.076149 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.076950 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.078251 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.078741 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.078917 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.079193 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.079219 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.081979 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.082334 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.083634 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.083760 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.084517 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.086201 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.086823 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.086922 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.086956 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.087672 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.088679 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.089299 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.091145 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.091817 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.092840 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.094509 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.095282 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.095279 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.095944 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.097771 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.098503 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.100089 4969 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.100337 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.101075 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.101891 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.103322 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.105251 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.106494 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.109692 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.110103 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.111570 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.112738 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.113623 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.114159 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.114208 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.114902 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.115952 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.116625 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.117594 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.118171 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.118992 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.119527 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.120356 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.121060 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.121896 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.122338 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.122810 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.123670 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.124217 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.125242 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.125632 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131149 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131193 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131239 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131248 4969 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131258 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131266 4969 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131275 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131283 4969 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131293 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131302 4969 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131313 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131322 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131333 4969 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131341 4969 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131349 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131357 4969 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131365 4969 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131373 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131380 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131389 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131398 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131408 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131416 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131436 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131444 4969 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131451 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131459 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131467 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131476 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131484 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131492 4969 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131501 4969 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131508 4969 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131518 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131525 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131533 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131541 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131550 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131558 4969 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131565 4969 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131574 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131582 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131590 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131599 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131607 4969 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131616 4969 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131624 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131632 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131640 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131649 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131658 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131666 4969 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131674 4969 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131682 4969 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131690 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131699 4969 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131706 4969 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131714 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131722 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131730 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131739 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131748 4969 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131757 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131765 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131773 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131781 4969 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131791 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131799 4969 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131808 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131817 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131825 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131844 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131858 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131866 4969 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131878 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131886 4969 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131914 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131937 4969 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131951 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131965 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131979 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.131992 4969 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132003 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132014 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132025 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132036 4969 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132048 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132059 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132070 4969 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132081 4969 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132101 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132113 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132130 4969 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132145 4969 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132163 4969 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132179 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132195 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132213 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132214 4969 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132239 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132247 4969 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132256 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132264 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132265 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132272 4969 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132558 4969 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132572 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132580 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132589 4969 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132598 4969 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132606 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132615 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132624 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132631 4969 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132640 4969 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132667 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132683 4969 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132695 4969 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132709 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132719 4969 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132731 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132744 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132756 4969 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132768 4969 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132780 4969 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132791 4969 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132802 4969 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132813 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132824 4969 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132835 4969 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132845 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132856 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132867 4969 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132877 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132888 4969 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132898 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132909 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132921 4969 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132931 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132942 4969 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132953 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132964 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132974 4969 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132985 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.132996 4969 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.133007 4969 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.133019 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.133033 4969 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.133043 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.133053 4969 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.133064 4969 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.133075 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.137174 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.148287 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.157025 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.169022 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.177058 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.197157 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.198662 4969 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183" exitCode=255 Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.198699 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183"} Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.209432 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.209816 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.209879 4969 scope.go:117] "RemoveContainer" containerID="fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.220407 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.230620 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.239720 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.249674 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.258921 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.290232 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.298794 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.303305 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 08:16:33 crc kubenswrapper[4969]: W1004 08:16:33.309042 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-7cd94788305748033228c533563d35322514a5551878349b61a0083b10782607 WatchSource:0}: Error finding container 7cd94788305748033228c533563d35322514a5551878349b61a0083b10782607: Status 404 returned error can't find the container with id 7cd94788305748033228c533563d35322514a5551878349b61a0083b10782607 Oct 04 08:16:33 crc kubenswrapper[4969]: W1004 08:16:33.319068 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-f4b6a3b7f995d48529220b88466139de33c2a1161632bb89ba61cef4cdcba297 WatchSource:0}: Error finding container f4b6a3b7f995d48529220b88466139de33c2a1161632bb89ba61cef4cdcba297: Status 404 returned error can't find the container with id f4b6a3b7f995d48529220b88466139de33c2a1161632bb89ba61cef4cdcba297 Oct 04 08:16:33 crc kubenswrapper[4969]: W1004 08:16:33.321518 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-c7cc9d446879fb2cef0ff39347e1878afa02bf63f2feb8f4ae02764ae012e1c8 WatchSource:0}: Error finding container c7cc9d446879fb2cef0ff39347e1878afa02bf63f2feb8f4ae02764ae012e1c8: Status 404 returned error can't find the container with id c7cc9d446879fb2cef0ff39347e1878afa02bf63f2feb8f4ae02764ae012e1c8 Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.535890 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.536033 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:16:34.535987126 +0000 UTC m=+22.290255950 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.636831 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.636878 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.636898 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:33 crc kubenswrapper[4969]: I1004 08:16:33.636917 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.636996 4969 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637029 4969 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637038 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637290 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637340 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637356 4969 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637081 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:34.637060895 +0000 UTC m=+22.391329709 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637469 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:34.637438754 +0000 UTC m=+22.391707568 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637457 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637485 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:34.637478845 +0000 UTC m=+22.391747659 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637506 4969 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:33 crc kubenswrapper[4969]: E1004 08:16:33.637584 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:34.637567957 +0000 UTC m=+22.391836791 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.054117 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.054242 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.145062 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-bc2w6"] Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.145438 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-ngkmm"] Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.145598 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.145709 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ngkmm" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.147311 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.147774 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.148227 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.148543 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.148776 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.149478 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.149893 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.150131 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.159906 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.170216 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.181599 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.195732 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.202825 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a"} Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.202891 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029"} Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.202908 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c7cc9d446879fb2cef0ff39347e1878afa02bf63f2feb8f4ae02764ae012e1c8"} Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.203883 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f4b6a3b7f995d48529220b88466139de33c2a1161632bb89ba61cef4cdcba297"} Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.205432 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f"} Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.205474 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7cd94788305748033228c533563d35322514a5551878349b61a0083b10782607"} Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.207939 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.209847 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc"} Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.210231 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.213440 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.229287 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.242015 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/69d1f843-03d6-403f-8ab3-796e2c97af4f-rootfs\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.242066 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wds2b\" (UniqueName: \"kubernetes.io/projected/16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f-kube-api-access-wds2b\") pod \"node-resolver-ngkmm\" (UID: \"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\") " pod="openshift-dns/node-resolver-ngkmm" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.242105 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f-hosts-file\") pod \"node-resolver-ngkmm\" (UID: \"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\") " pod="openshift-dns/node-resolver-ngkmm" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.242136 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.242280 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/69d1f843-03d6-403f-8ab3-796e2c97af4f-mcd-auth-proxy-config\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.242350 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/69d1f843-03d6-403f-8ab3-796e2c97af4f-proxy-tls\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.242400 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f75xj\" (UniqueName: \"kubernetes.io/projected/69d1f843-03d6-403f-8ab3-796e2c97af4f-kube-api-access-f75xj\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.255380 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.270391 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.285102 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.306169 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.322663 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.334760 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.342815 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/69d1f843-03d6-403f-8ab3-796e2c97af4f-rootfs\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.342857 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wds2b\" (UniqueName: \"kubernetes.io/projected/16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f-kube-api-access-wds2b\") pod \"node-resolver-ngkmm\" (UID: \"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\") " pod="openshift-dns/node-resolver-ngkmm" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.342962 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f-hosts-file\") pod \"node-resolver-ngkmm\" (UID: \"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\") " pod="openshift-dns/node-resolver-ngkmm" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.342997 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/69d1f843-03d6-403f-8ab3-796e2c97af4f-rootfs\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.343259 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f-hosts-file\") pod \"node-resolver-ngkmm\" (UID: \"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\") " pod="openshift-dns/node-resolver-ngkmm" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.343012 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/69d1f843-03d6-403f-8ab3-796e2c97af4f-mcd-auth-proxy-config\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.344244 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/69d1f843-03d6-403f-8ab3-796e2c97af4f-proxy-tls\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.344370 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/69d1f843-03d6-403f-8ab3-796e2c97af4f-mcd-auth-proxy-config\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.344398 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f75xj\" (UniqueName: \"kubernetes.io/projected/69d1f843-03d6-403f-8ab3-796e2c97af4f-kube-api-access-f75xj\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.351925 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.365727 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.379752 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.392411 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.406554 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.446853 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/69d1f843-03d6-403f-8ab3-796e2c97af4f-proxy-tls\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.446918 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wds2b\" (UniqueName: \"kubernetes.io/projected/16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f-kube-api-access-wds2b\") pod \"node-resolver-ngkmm\" (UID: \"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\") " pod="openshift-dns/node-resolver-ngkmm" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.447116 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f75xj\" (UniqueName: \"kubernetes.io/projected/69d1f843-03d6-403f-8ab3-796e2c97af4f-kube-api-access-f75xj\") pod \"machine-config-daemon-bc2w6\" (UID: \"69d1f843-03d6-403f-8ab3-796e2c97af4f\") " pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.459167 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ngkmm" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.466150 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.522203 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-2tw2d"] Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.522904 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dl8q8"] Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.523127 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.523793 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.524647 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-mf8dr"] Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.525020 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.530766 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.530966 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.531109 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.531238 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.531330 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.531518 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.531683 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.531781 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.531925 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.532025 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.538371 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.538685 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.538765 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.538983 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.546502 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.546711 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:16:36.546688434 +0000 UTC m=+24.300957248 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.551887 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.566676 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.579278 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.599169 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.610442 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.624581 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647615 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-etc-openvswitch\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647658 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-log-socket\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647679 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-bin\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647698 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-netd\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647736 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647772 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-env-overrides\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647806 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flkkk\" (UniqueName: \"kubernetes.io/projected/5db532e6-047f-472c-92d8-ff80bef1bd19-kube-api-access-flkkk\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647884 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-os-release\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647922 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-hostroot\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647942 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-var-lib-openvswitch\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647965 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-systemd\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.647982 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-openvswitch\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648011 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648026 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-var-lib-kubelet\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648046 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-cnibin\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648063 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-conf-dir\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.648085 4969 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.648137 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:36.648122553 +0000 UTC m=+24.402391367 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648151 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-run-multus-certs\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648167 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-ovn-kubernetes\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648183 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-os-release\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648205 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.648259 4969 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648269 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-run-netns\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.648284 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:36.648275956 +0000 UTC m=+24.402544770 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648313 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-kubelet\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648337 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-slash\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648369 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648468 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-netns\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.648503 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.648520 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.648530 4969 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648505 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5db532e6-047f-472c-92d8-ff80bef1bd19-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.648594 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:36.648579443 +0000 UTC m=+24.402848257 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648612 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-daemon-config\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648630 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648645 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-system-cni-dir\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648663 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-etc-kubernetes\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648677 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovn-node-metrics-cert\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648694 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-var-lib-cni-bin\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648710 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-ovn\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648723 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-config\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648740 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-var-lib-cni-multus\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648755 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc784\" (UniqueName: \"kubernetes.io/projected/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-kube-api-access-rc784\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648782 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-systemd-units\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648821 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-cni-dir\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648864 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-cni-binary-copy\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648899 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-run-k8s-cni-cncf-io\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648937 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-node-log\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648957 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-script-lib\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648971 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcv2d\" (UniqueName: \"kubernetes.io/projected/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-kube-api-access-bcv2d\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.648986 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-system-cni-dir\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.649017 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-cnibin\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.649067 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.649096 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-socket-dir-parent\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.649133 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5db532e6-047f-472c-92d8-ff80bef1bd19-cni-binary-copy\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.649167 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.649184 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.649196 4969 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:34 crc kubenswrapper[4969]: E1004 08:16:34.649227 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:36.649217309 +0000 UTC m=+24.403486243 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.667018 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.697061 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.707927 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.723719 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.735300 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.746119 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749771 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-daemon-config\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749804 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749821 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-system-cni-dir\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749838 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-etc-kubernetes\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749858 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovn-node-metrics-cert\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749875 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-var-lib-cni-bin\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749890 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-ovn\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749907 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-config\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749934 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-etc-kubernetes\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749988 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-var-lib-cni-bin\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750066 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-var-lib-cni-multus\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750097 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-system-cni-dir\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750177 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-ovn\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.749923 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-var-lib-cni-multus\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750514 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc784\" (UniqueName: \"kubernetes.io/projected/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-kube-api-access-rc784\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750533 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-systemd-units\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750553 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-cni-dir\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750567 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-cni-binary-copy\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750568 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-systemd-units\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750582 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-run-k8s-cni-cncf-io\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750572 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-config\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750597 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-node-log\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750614 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-script-lib\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750628 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-daemon-config\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750643 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-cni-dir\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750662 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-cnibin\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750652 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-node-log\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750636 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-cnibin\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750697 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcv2d\" (UniqueName: \"kubernetes.io/projected/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-kube-api-access-bcv2d\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750632 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-run-k8s-cni-cncf-io\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750765 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-system-cni-dir\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750786 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-socket-dir-parent\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750830 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-system-cni-dir\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750841 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-socket-dir-parent\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750868 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5db532e6-047f-472c-92d8-ff80bef1bd19-cni-binary-copy\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750898 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-bin\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.750960 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-cni-binary-copy\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751006 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-bin\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751099 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-script-lib\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751503 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5db532e6-047f-472c-92d8-ff80bef1bd19-cni-binary-copy\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751557 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-netd\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751576 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751591 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-env-overrides\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751964 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flkkk\" (UniqueName: \"kubernetes.io/projected/5db532e6-047f-472c-92d8-ff80bef1bd19-kube-api-access-flkkk\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751991 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-etc-openvswitch\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752008 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-log-socket\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752023 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-os-release\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752040 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-hostroot\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752057 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-var-lib-openvswitch\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752073 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-openvswitch\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752087 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-systemd\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752109 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-var-lib-kubelet\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752125 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-cnibin\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752139 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-conf-dir\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752156 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-run-multus-certs\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752191 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-ovn-kubernetes\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752212 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-os-release\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752231 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-run-netns\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752247 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-kubelet\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751941 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-env-overrides\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751631 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752292 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-openvswitch\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752330 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-hostroot\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752335 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-log-socket\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752354 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-var-lib-openvswitch\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.751659 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-netd\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752391 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-run-multus-certs\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752437 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-systemd\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752471 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-var-lib-kubelet\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752509 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-cnibin\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752515 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-os-release\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752525 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-ovn-kubernetes\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752539 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-multus-conf-dir\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752542 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-host-run-netns\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752567 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-etc-openvswitch\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752568 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-kubelet\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752659 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-slash\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752705 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-os-release\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.754773 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovn-node-metrics-cert\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.752263 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-slash\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.757662 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-netns\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.757685 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5db532e6-047f-472c-92d8-ff80bef1bd19-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.757740 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-netns\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.758224 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5db532e6-047f-472c-92d8-ff80bef1bd19-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.759098 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5db532e6-047f-472c-92d8-ff80bef1bd19-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.764481 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.769853 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc784\" (UniqueName: \"kubernetes.io/projected/8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6-kube-api-access-rc784\") pod \"multus-mf8dr\" (UID: \"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\") " pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.772190 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flkkk\" (UniqueName: \"kubernetes.io/projected/5db532e6-047f-472c-92d8-ff80bef1bd19-kube-api-access-flkkk\") pod \"multus-additional-cni-plugins-2tw2d\" (UID: \"5db532e6-047f-472c-92d8-ff80bef1bd19\") " pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.774804 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcv2d\" (UniqueName: \"kubernetes.io/projected/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-kube-api-access-bcv2d\") pod \"ovnkube-node-dl8q8\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.777377 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.788480 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.796711 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.808533 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.822227 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.835519 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.843799 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.848004 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: W1004 08:16:34.853732 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5db532e6_047f_472c_92d8_ff80bef1bd19.slice/crio-20e953daa13b0ef4f8eed18e490620496a9b60d5c7d3359ea18546d2d7b9b54c WatchSource:0}: Error finding container 20e953daa13b0ef4f8eed18e490620496a9b60d5c7d3359ea18546d2d7b9b54c: Status 404 returned error can't find the container with id 20e953daa13b0ef4f8eed18e490620496a9b60d5c7d3359ea18546d2d7b9b54c Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.860682 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.862920 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.865993 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-mf8dr" Oct 04 08:16:34 crc kubenswrapper[4969]: I1004 08:16:34.876129 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:34Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:34 crc kubenswrapper[4969]: W1004 08:16:34.877586 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e5258c5_cab0_4d0b_b5fd_0e8ffd1f19be.slice/crio-91fea041a813870b1ef1f48eca51b8092fe25022c8e2e2d89b8f7463e1519b08 WatchSource:0}: Error finding container 91fea041a813870b1ef1f48eca51b8092fe25022c8e2e2d89b8f7463e1519b08: Status 404 returned error can't find the container with id 91fea041a813870b1ef1f48eca51b8092fe25022c8e2e2d89b8f7463e1519b08 Oct 04 08:16:34 crc kubenswrapper[4969]: W1004 08:16:34.881657 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f4f1158_78b1_4759_b3b1_a0cfd8ee1ba6.slice/crio-5a906ab1787e1fa007c26517f2cb2bc5c22b3bb6beeabefef2c9933a0b666adc WatchSource:0}: Error finding container 5a906ab1787e1fa007c26517f2cb2bc5c22b3bb6beeabefef2c9933a0b666adc: Status 404 returned error can't find the container with id 5a906ab1787e1fa007c26517f2cb2bc5c22b3bb6beeabefef2c9933a0b666adc Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.054288 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.054310 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:35 crc kubenswrapper[4969]: E1004 08:16:35.054400 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:35 crc kubenswrapper[4969]: E1004 08:16:35.054725 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.058805 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.059723 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.060356 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.061023 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.062440 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.063171 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.064156 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.064733 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.065844 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.214365 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d" exitCode=0 Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.214454 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.214516 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"91fea041a813870b1ef1f48eca51b8092fe25022c8e2e2d89b8f7463e1519b08"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.216755 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.216793 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.216808 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"ae28aed03133a823b570e114d69ef204544ab07205383e399f13df11d55ea29b"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.218380 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mf8dr" event={"ID":"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6","Type":"ContainerStarted","Data":"fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.218429 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mf8dr" event={"ID":"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6","Type":"ContainerStarted","Data":"5a906ab1787e1fa007c26517f2cb2bc5c22b3bb6beeabefef2c9933a0b666adc"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.220748 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerStarted","Data":"a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.220806 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerStarted","Data":"20e953daa13b0ef4f8eed18e490620496a9b60d5c7d3359ea18546d2d7b9b54c"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.222871 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ngkmm" event={"ID":"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f","Type":"ContainerStarted","Data":"381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.222913 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ngkmm" event={"ID":"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f","Type":"ContainerStarted","Data":"676cb2cf40ca3672ce5c46de8fb6283fcb889585ca77ee7e98df752caa760e09"} Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.232897 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.254398 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.267077 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.283332 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.296734 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.309178 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.319553 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.334140 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.347645 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.361721 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.372980 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.394130 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.411273 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.422962 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.433050 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.446025 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.457716 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.470917 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.484102 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.495848 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.509252 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.520686 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.532109 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:35 crc kubenswrapper[4969]: I1004 08:16:35.550069 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:35Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.054735 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.054848 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.093753 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.097518 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.101961 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.104532 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.117320 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.128003 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.144035 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.161622 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.175568 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.194163 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.210863 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.227335 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2"} Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.227570 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.236981 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40"} Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.237033 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a"} Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.237047 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb"} Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.237060 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b"} Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.237071 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931"} Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.237082 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68"} Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.239140 4969 generic.go:334] "Generic (PLEG): container finished" podID="5db532e6-047f-472c-92d8-ff80bef1bd19" containerID="a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625" exitCode=0 Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.239198 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerDied","Data":"a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625"} Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.247903 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.267821 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.283507 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.294917 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.315491 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.327180 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.342290 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.354545 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-gz7t7"] Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.354872 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.358955 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.359035 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.359149 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.359232 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.359233 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.377163 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.386714 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.401911 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.413742 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.432823 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.444082 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.456286 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.466767 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.472733 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb8l7\" (UniqueName: \"kubernetes.io/projected/16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff-kube-api-access-fb8l7\") pod \"node-ca-gz7t7\" (UID: \"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\") " pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.472926 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff-host\") pod \"node-ca-gz7t7\" (UID: \"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\") " pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.472977 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff-serviceca\") pod \"node-ca-gz7t7\" (UID: \"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\") " pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.484966 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.503109 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.516522 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.530540 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.558008 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.573796 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.573906 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff-host\") pod \"node-ca-gz7t7\" (UID: \"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\") " pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.573970 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:16:40.573950022 +0000 UTC m=+28.328218846 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.573995 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff-serviceca\") pod \"node-ca-gz7t7\" (UID: \"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\") " pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.574020 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff-host\") pod \"node-ca-gz7t7\" (UID: \"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\") " pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.574055 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb8l7\" (UniqueName: \"kubernetes.io/projected/16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff-kube-api-access-fb8l7\") pod \"node-ca-gz7t7\" (UID: \"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\") " pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.575466 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff-serviceca\") pod \"node-ca-gz7t7\" (UID: \"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\") " pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.597768 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.631150 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb8l7\" (UniqueName: \"kubernetes.io/projected/16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff-kube-api-access-fb8l7\") pod \"node-ca-gz7t7\" (UID: \"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\") " pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.659894 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.675898 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.675991 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.676068 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.676132 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.676273 4969 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.676368 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:40.676336874 +0000 UTC m=+28.430605738 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677053 4969 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677147 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:40.677118423 +0000 UTC m=+28.431387277 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677280 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677311 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677334 4969 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677399 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:40.677377559 +0000 UTC m=+28.431646423 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677549 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677580 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677600 4969 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:36 crc kubenswrapper[4969]: E1004 08:16:36.677838 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:40.677817099 +0000 UTC m=+28.432085953 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.705870 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.731053 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gz7t7" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.737639 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: W1004 08:16:36.742974 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16b3e242_ff6c_4c8b_aeb2_aed31bfc5eff.slice/crio-e602eeeaba0b3aba1ed77ef5dabf4646c367af3a62c7d8628e25318ba99e9c8e WatchSource:0}: Error finding container e602eeeaba0b3aba1ed77ef5dabf4646c367af3a62c7d8628e25318ba99e9c8e: Status 404 returned error can't find the container with id e602eeeaba0b3aba1ed77ef5dabf4646c367af3a62c7d8628e25318ba99e9c8e Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.778531 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.816958 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.859908 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.896723 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:36 crc kubenswrapper[4969]: I1004 08:16:36.933839 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:36Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.055819 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:37 crc kubenswrapper[4969]: E1004 08:16:37.056341 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.056949 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:37 crc kubenswrapper[4969]: E1004 08:16:37.057047 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.243383 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerStarted","Data":"5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113"} Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.244721 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gz7t7" event={"ID":"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff","Type":"ContainerStarted","Data":"dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d"} Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.244774 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gz7t7" event={"ID":"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff","Type":"ContainerStarted","Data":"e602eeeaba0b3aba1ed77ef5dabf4646c367af3a62c7d8628e25318ba99e9c8e"} Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.258167 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.272121 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.284224 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.295954 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.311311 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.324274 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.338806 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.361205 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.377175 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.394319 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.409140 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.425194 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.462005 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.498487 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.539663 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.581217 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.619898 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.663257 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.695976 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.749755 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.781088 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.821752 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.862892 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.896971 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.946095 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:37 crc kubenswrapper[4969]: I1004 08:16:37.982847 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:37Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.016603 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.054175 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:38 crc kubenswrapper[4969]: E1004 08:16:38.054408 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.062287 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.249735 4969 generic.go:334] "Generic (PLEG): container finished" podID="5db532e6-047f-472c-92d8-ff80bef1bd19" containerID="5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113" exitCode=0 Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.249807 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerDied","Data":"5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113"} Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.276019 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.297707 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.311614 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.338733 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.355053 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.379048 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.395629 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.419698 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.439689 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.456852 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.497526 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.538570 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.579721 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.615783 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.919067 4969 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.922047 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.922107 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.922125 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.922240 4969 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.933010 4969 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.933471 4969 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.935211 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.935266 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.935284 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.935307 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.935325 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:38Z","lastTransitionTime":"2025-10-04T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:38 crc kubenswrapper[4969]: E1004 08:16:38.956542 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.961368 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.961465 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.961491 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.961522 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.961541 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:38Z","lastTransitionTime":"2025-10-04T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:38 crc kubenswrapper[4969]: E1004 08:16:38.986197 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:38Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.992378 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.992474 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.992496 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.992521 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:38 crc kubenswrapper[4969]: I1004 08:16:38.992539 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:38Z","lastTransitionTime":"2025-10-04T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: E1004 08:16:39.011138 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.016107 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.016172 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.016196 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.016223 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.016243 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: E1004 08:16:39.035296 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.040275 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.040389 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.040416 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.040485 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.040509 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.054351 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.054489 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:39 crc kubenswrapper[4969]: E1004 08:16:39.054652 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:39 crc kubenswrapper[4969]: E1004 08:16:39.055349 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:39 crc kubenswrapper[4969]: E1004 08:16:39.065673 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: E1004 08:16:39.065926 4969 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.068197 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.068257 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.068278 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.068304 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.068323 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.171604 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.171651 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.171661 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.171677 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.171688 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.258160 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.261005 4969 generic.go:334] "Generic (PLEG): container finished" podID="5db532e6-047f-472c-92d8-ff80bef1bd19" containerID="bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f" exitCode=0 Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.261047 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerDied","Data":"bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.274115 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.274163 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.274174 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.274193 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.274206 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.283337 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.299314 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.319314 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.340665 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.355845 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.374001 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.376895 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.376928 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.376941 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.376956 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.376965 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.389862 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.404543 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.423246 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.438323 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.456047 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.471874 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.479275 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.479325 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.479339 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.479361 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.479376 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.485950 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.500450 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:39Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.582252 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.582308 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.582325 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.582348 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.582363 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.684685 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.684740 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.684756 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.684777 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.684798 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.787398 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.787494 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.787512 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.787536 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.787553 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.890158 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.890217 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.890236 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.890262 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.890281 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.993574 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.993639 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.993660 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.993688 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:39 crc kubenswrapper[4969]: I1004 08:16:39.993707 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:39Z","lastTransitionTime":"2025-10-04T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.054706 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.054890 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.096377 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.096469 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.096488 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.096512 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.096531 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:40Z","lastTransitionTime":"2025-10-04T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.199901 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.199952 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.199963 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.199982 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.199995 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:40Z","lastTransitionTime":"2025-10-04T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.268337 4969 generic.go:334] "Generic (PLEG): container finished" podID="5db532e6-047f-472c-92d8-ff80bef1bd19" containerID="dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca" exitCode=0 Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.268474 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerDied","Data":"dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.288392 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.302830 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.302906 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.302930 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.302961 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.302986 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:40Z","lastTransitionTime":"2025-10-04T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.307837 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.334794 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.356979 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.376255 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.391761 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.405886 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.405948 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.405967 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.405995 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.406016 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:40Z","lastTransitionTime":"2025-10-04T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.413212 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.433072 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.453385 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.475887 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.493690 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.509365 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.509678 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.509705 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.509715 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.509733 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.509745 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:40Z","lastTransitionTime":"2025-10-04T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.524574 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.539862 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:40Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.613039 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.613112 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.613136 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.613183 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.613209 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:40Z","lastTransitionTime":"2025-10-04T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.619538 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.619837 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:16:48.619809262 +0000 UTC m=+36.374078116 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.716554 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.716631 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.716652 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.716681 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.716702 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:40Z","lastTransitionTime":"2025-10-04T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.720405 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.720513 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.720555 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.720593 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720661 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720707 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720704 4969 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720731 4969 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720761 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720763 4969 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720784 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720808 4969 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720808 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:48.720783219 +0000 UTC m=+36.475052063 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720880 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:48.720861901 +0000 UTC m=+36.475130755 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720901 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:48.720890132 +0000 UTC m=+36.475158976 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:40 crc kubenswrapper[4969]: E1004 08:16:40.720924 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:48.720914002 +0000 UTC m=+36.475182856 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.820560 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.820624 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.820641 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.820669 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.820695 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:40Z","lastTransitionTime":"2025-10-04T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.923751 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.923824 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.923863 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.923893 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:40 crc kubenswrapper[4969]: I1004 08:16:40.923913 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:40Z","lastTransitionTime":"2025-10-04T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.026853 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.027378 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.027404 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.027470 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.027495 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.054782 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.054905 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:41 crc kubenswrapper[4969]: E1004 08:16:41.055070 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:41 crc kubenswrapper[4969]: E1004 08:16:41.055274 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.130958 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.131005 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.131020 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.131047 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.131062 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.233343 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.233392 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.233404 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.233446 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.233459 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.278196 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.278642 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.278664 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.284127 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerStarted","Data":"a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.297266 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.314680 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.315027 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.318718 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.333072 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.335956 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.335986 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.335998 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.336016 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.336026 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.347268 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.362455 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.374584 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.391388 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.401787 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.417275 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.427901 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.438749 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.438783 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.438791 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.438805 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.438813 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.438960 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.452662 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.462960 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.484541 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.504130 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.517056 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.528179 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.541736 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.541789 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.541803 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.541820 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.541833 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.547123 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.563162 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.574369 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.586409 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.597187 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.608139 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.620015 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.641000 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.643272 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.643306 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.643315 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.643327 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.643337 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.652992 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.663518 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.683013 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.746018 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.746089 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.746108 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.746132 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.746156 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.849744 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.849815 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.849836 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.849862 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.849880 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.952748 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.952782 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.952792 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.952810 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:41 crc kubenswrapper[4969]: I1004 08:16:41.952821 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:41Z","lastTransitionTime":"2025-10-04T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.054543 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:42 crc kubenswrapper[4969]: E1004 08:16:42.054742 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.055314 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.055369 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.055391 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.055455 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.055482 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.158554 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.158634 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.158660 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.158687 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.158705 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.262402 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.262517 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.262545 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.262578 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.262603 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.292155 4969 generic.go:334] "Generic (PLEG): container finished" podID="5db532e6-047f-472c-92d8-ff80bef1bd19" containerID="a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd" exitCode=0 Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.292238 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerDied","Data":"a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.292405 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.315729 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.341795 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.354924 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.365575 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.365608 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.365618 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.365635 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.365645 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.370882 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.394832 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.410495 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.428279 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.439766 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.457290 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.468257 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.468312 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.468330 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.468351 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.468365 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.470598 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.481974 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.497957 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.515171 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.528161 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:42Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.571290 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.571327 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.571336 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.571350 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.571359 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.674676 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.674765 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.674790 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.674826 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.674850 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.777248 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.777296 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.777308 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.777326 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.777339 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.880654 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.880995 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.881006 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.881022 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.881035 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.984389 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.984473 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.984505 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.984539 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:42 crc kubenswrapper[4969]: I1004 08:16:42.984559 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:42Z","lastTransitionTime":"2025-10-04T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.056645 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.056809 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:43 crc kubenswrapper[4969]: E1004 08:16:43.056877 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:43 crc kubenswrapper[4969]: E1004 08:16:43.057043 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.074227 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.092051 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.092341 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.092468 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.092585 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.092674 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:43Z","lastTransitionTime":"2025-10-04T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.094949 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.112001 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.135598 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.155738 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.184152 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.194588 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.194785 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.194916 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.195043 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.195164 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:43Z","lastTransitionTime":"2025-10-04T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.199723 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.213588 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.226966 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.240785 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.258923 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.271729 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.284954 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.296912 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.296951 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.296967 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.296992 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.297010 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:43Z","lastTransitionTime":"2025-10-04T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.298172 4969 generic.go:334] "Generic (PLEG): container finished" podID="5db532e6-047f-472c-92d8-ff80bef1bd19" containerID="8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c" exitCode=0 Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.298245 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerDied","Data":"8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.298390 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.299120 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.314878 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.329785 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.339604 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.361287 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.375705 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.389873 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.401246 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.401302 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.401320 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.401345 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.401366 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:43Z","lastTransitionTime":"2025-10-04T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.407291 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.426527 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.445776 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.465939 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.481300 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.496644 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.503292 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.503316 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.503324 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.503336 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.503344 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:43Z","lastTransitionTime":"2025-10-04T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.509803 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.528234 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.604990 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.605030 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.605038 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.605053 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.605063 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:43Z","lastTransitionTime":"2025-10-04T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.708954 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.709009 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.709026 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.709049 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.709067 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:43Z","lastTransitionTime":"2025-10-04T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.812500 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.812549 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.812565 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.812583 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.812594 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:43Z","lastTransitionTime":"2025-10-04T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.915658 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.915713 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.915730 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.915754 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:43 crc kubenswrapper[4969]: I1004 08:16:43.915771 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:43Z","lastTransitionTime":"2025-10-04T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.018321 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.018385 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.018402 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.018466 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.018494 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.054716 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:44 crc kubenswrapper[4969]: E1004 08:16:44.054860 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.121141 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.121179 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.121189 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.121203 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.121215 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.223964 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.224001 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.224012 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.224029 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.224040 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.278662 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.305707 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" event={"ID":"5db532e6-047f-472c-92d8-ff80bef1bd19","Type":"ContainerStarted","Data":"e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.322400 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.326147 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.326189 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.326200 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.326220 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.326232 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.342284 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.362062 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.380329 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.398081 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.417527 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.429169 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.429227 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.429244 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.429269 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.429290 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.436141 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.464188 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.480517 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.493770 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.505925 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.523485 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.532115 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.532170 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.532188 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.532212 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.532230 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.540367 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.553785 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:44Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.636211 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.636264 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.636283 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.636306 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.636324 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.739676 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.739744 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.739762 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.739786 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.739804 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.843090 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.843147 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.843165 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.843190 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.843208 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.946683 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.946739 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.946755 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.946780 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:44 crc kubenswrapper[4969]: I1004 08:16:44.946797 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:44Z","lastTransitionTime":"2025-10-04T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.049866 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.049933 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.049957 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.049988 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.050012 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.054527 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.054622 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:45 crc kubenswrapper[4969]: E1004 08:16:45.054729 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:45 crc kubenswrapper[4969]: E1004 08:16:45.055011 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.153559 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.153630 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.153654 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.153683 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.153706 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.256756 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.256814 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.256833 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.256855 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.256872 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.312482 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/0.log" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.317124 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4" exitCode=1 Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.317201 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.318375 4969 scope.go:117] "RemoveContainer" containerID="00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.341591 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.359736 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.360200 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.360226 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.360257 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.360288 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.363068 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.380817 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.405707 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.426765 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.448636 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.463099 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.463159 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.463178 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.463203 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.463221 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.470036 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.503251 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"message\\\":\\\"client-go/informers/factory.go:160\\\\nI1004 08:16:44.412151 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 08:16:44.412204 6233 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:16:44.412789 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 08:16:44.412877 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 08:16:44.412895 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 08:16:44.412916 6233 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 08:16:44.412930 6233 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 08:16:44.412964 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:16:44.412991 6233 factory.go:656] Stopping watch factory\\\\nI1004 08:16:44.413007 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:16:44.413036 6233 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 08:16:44.413055 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 08:16:44.413067 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 08:16:44.413079 6233 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 08:16:44.413090 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 08:16:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.521025 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.545468 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.567578 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.567631 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.567649 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.567672 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.567687 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.570625 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.588674 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.607724 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.629210 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:45Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.670582 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.670631 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.670641 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.670659 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.670671 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.774136 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.774214 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.774237 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.774266 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.774290 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.876749 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.876792 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.876804 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.876822 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.876833 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.979203 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.979244 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.979259 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.979277 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:45 crc kubenswrapper[4969]: I1004 08:16:45.979287 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:45Z","lastTransitionTime":"2025-10-04T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.055217 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:46 crc kubenswrapper[4969]: E1004 08:16:46.055534 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.081749 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.081791 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.081804 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.081820 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.081833 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:46Z","lastTransitionTime":"2025-10-04T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.185950 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.186003 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.186020 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.186042 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.186059 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:46Z","lastTransitionTime":"2025-10-04T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.288244 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.288297 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.288315 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.288337 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.288353 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:46Z","lastTransitionTime":"2025-10-04T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.324904 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/0.log" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.329464 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.330380 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.351393 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.372025 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.391760 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.392592 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.392653 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.392678 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.392701 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.392718 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:46Z","lastTransitionTime":"2025-10-04T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.411744 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.432639 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.454663 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.471667 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.495220 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.495288 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.495313 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.495341 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.495364 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:46Z","lastTransitionTime":"2025-10-04T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.502123 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"message\\\":\\\"client-go/informers/factory.go:160\\\\nI1004 08:16:44.412151 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 08:16:44.412204 6233 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:16:44.412789 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 08:16:44.412877 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 08:16:44.412895 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 08:16:44.412916 6233 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 08:16:44.412930 6233 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 08:16:44.412964 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:16:44.412991 6233 factory.go:656] Stopping watch factory\\\\nI1004 08:16:44.413007 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:16:44.413036 6233 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 08:16:44.413055 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 08:16:44.413067 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 08:16:44.413079 6233 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 08:16:44.413090 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 08:16:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.518062 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.539381 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.558153 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.574365 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.598303 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.598384 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.598403 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.598462 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.598484 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:46Z","lastTransitionTime":"2025-10-04T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.600135 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.620764 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.701393 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.701527 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.701554 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.701587 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.701612 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:46Z","lastTransitionTime":"2025-10-04T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.804902 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.804979 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.805025 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.805056 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.805078 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:46Z","lastTransitionTime":"2025-10-04T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.907897 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.907948 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.907966 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.907991 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:46 crc kubenswrapper[4969]: I1004 08:16:46.908008 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:46Z","lastTransitionTime":"2025-10-04T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.011285 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.011356 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.011379 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.011408 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.011462 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.054890 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.054928 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:47 crc kubenswrapper[4969]: E1004 08:16:47.055072 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:47 crc kubenswrapper[4969]: E1004 08:16:47.055214 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.128951 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t"] Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.130082 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.131280 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.131331 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.131358 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.131385 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.131402 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.132262 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.132633 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.151809 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.167121 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.182856 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.200541 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.215959 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.229018 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.233767 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.233853 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.233878 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.233907 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.233931 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.245026 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.263627 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.291617 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpzbw\" (UniqueName: \"kubernetes.io/projected/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-kube-api-access-zpzbw\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.291700 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.291750 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.292115 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.295834 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"message\\\":\\\"client-go/informers/factory.go:160\\\\nI1004 08:16:44.412151 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 08:16:44.412204 6233 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:16:44.412789 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 08:16:44.412877 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 08:16:44.412895 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 08:16:44.412916 6233 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 08:16:44.412930 6233 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 08:16:44.412964 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:16:44.412991 6233 factory.go:656] Stopping watch factory\\\\nI1004 08:16:44.413007 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:16:44.413036 6233 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 08:16:44.413055 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 08:16:44.413067 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 08:16:44.413079 6233 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 08:16:44.413090 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 08:16:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.311048 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.335499 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.336094 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/1.log" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.336202 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.336246 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.336267 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.336295 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.336318 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.337252 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/0.log" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.340891 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165" exitCode=1 Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.340944 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.341097 4969 scope.go:117] "RemoveContainer" containerID="00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.342062 4969 scope.go:117] "RemoveContainer" containerID="dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165" Oct 04 08:16:47 crc kubenswrapper[4969]: E1004 08:16:47.342552 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.353622 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.369483 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.389111 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.393364 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.393515 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpzbw\" (UniqueName: \"kubernetes.io/projected/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-kube-api-access-zpzbw\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.393611 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.393673 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.394123 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.394150 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.399797 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.406516 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.416167 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpzbw\" (UniqueName: \"kubernetes.io/projected/ad39e34e-f40b-46e1-bf57-544dd3cd15ab-kube-api-access-zpzbw\") pod \"ovnkube-control-plane-749d76644c-fgf4t\" (UID: \"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.424668 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.438593 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.438654 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.438672 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.438695 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.438710 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.446328 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.452514 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.464934 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: W1004 08:16:47.467187 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad39e34e_f40b_46e1_bf57_544dd3cd15ab.slice/crio-3d37e34f43c17c7c09ed3cfc7c57bd7bc3fbf8a13970eac307293ef72393b1bd WatchSource:0}: Error finding container 3d37e34f43c17c7c09ed3cfc7c57bd7bc3fbf8a13970eac307293ef72393b1bd: Status 404 returned error can't find the container with id 3d37e34f43c17c7c09ed3cfc7c57bd7bc3fbf8a13970eac307293ef72393b1bd Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.483611 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.503277 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.519908 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.538884 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a0621595ae9ddb23d053df932599831b4433c511cdc1250e4e611572ff0cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"message\\\":\\\"client-go/informers/factory.go:160\\\\nI1004 08:16:44.412151 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 08:16:44.412204 6233 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:16:44.412789 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 08:16:44.412877 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 08:16:44.412895 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 08:16:44.412916 6233 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 08:16:44.412930 6233 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 08:16:44.412964 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:16:44.412991 6233 factory.go:656] Stopping watch factory\\\\nI1004 08:16:44.413007 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:16:44.413036 6233 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 08:16:44.413055 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 08:16:44.413067 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 08:16:44.413079 6233 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 08:16:44.413090 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 08:16:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:46Z\\\",\\\"message\\\":\\\" tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z]\\\\nI1004 08:16:46.256960 6427 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 08:16:46.257019 6427 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_router_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:169.254.0.2:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.540509 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.540548 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.540560 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.540575 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.540586 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.560860 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.575911 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.591597 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.603001 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.619223 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.633197 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.643364 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.643402 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.643413 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.643469 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.643484 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.645988 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.664875 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:47Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.747137 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.747221 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.747246 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.747273 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.747291 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.850487 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.850562 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.850580 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.850607 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.850626 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.953481 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.953538 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.953550 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.953568 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:47 crc kubenswrapper[4969]: I1004 08:16:47.953579 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:47Z","lastTransitionTime":"2025-10-04T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.055065 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.055218 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.057379 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.057498 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.057516 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.057537 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.057551 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.159492 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.159793 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.159920 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.160047 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.160202 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.262873 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.263171 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.263308 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.263501 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.263684 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.347509 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/1.log" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.353735 4969 scope.go:117] "RemoveContainer" containerID="dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165" Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.354043 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.355632 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" event={"ID":"ad39e34e-f40b-46e1-bf57-544dd3cd15ab","Type":"ContainerStarted","Data":"44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.355791 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" event={"ID":"ad39e34e-f40b-46e1-bf57-544dd3cd15ab","Type":"ContainerStarted","Data":"3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.355812 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" event={"ID":"ad39e34e-f40b-46e1-bf57-544dd3cd15ab","Type":"ContainerStarted","Data":"3d37e34f43c17c7c09ed3cfc7c57bd7bc3fbf8a13970eac307293ef72393b1bd"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.367357 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.367409 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.367450 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.367473 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.367489 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.382522 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.404224 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.420570 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.434815 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.449580 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.458311 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.469694 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.469721 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.469730 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.469743 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.469755 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.470828 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.483067 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.494198 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.511796 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.524674 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.534704 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.545250 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.556125 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.571858 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.571898 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.571910 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.571926 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.571939 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.574142 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:46Z\\\",\\\"message\\\":\\\" tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z]\\\\nI1004 08:16:46.256960 6427 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 08:16:46.257019 6427 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_router_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:169.254.0.2:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.590003 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.606353 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.619063 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.635938 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-kwwvg"] Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.636584 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.636642 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.638478 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.655657 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.672108 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.674167 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.674248 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.674275 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.674308 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.674332 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.692787 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.708069 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.708199 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-955nm\" (UniqueName: \"kubernetes.io/projected/01650d78-a230-4e2d-80bb-817099bc34f2-kube-api-access-955nm\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.708276 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:17:04.708247351 +0000 UTC m=+52.462516155 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.708440 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.709068 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.742191 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:46Z\\\",\\\"message\\\":\\\" tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z]\\\\nI1004 08:16:46.256960 6427 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 08:16:46.257019 6427 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_router_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:169.254.0.2:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.760869 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.772693 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.776573 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.776620 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.776638 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.776662 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.776679 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.788123 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.803832 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.810101 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.810193 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810241 4969 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.810250 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810294 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:04.810278964 +0000 UTC m=+52.564547778 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.810321 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.810383 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-955nm\" (UniqueName: \"kubernetes.io/projected/01650d78-a230-4e2d-80bb-817099bc34f2-kube-api-access-955nm\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.810405 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810488 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810508 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810521 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810527 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810555 4969 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810531 4969 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810600 4969 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810700 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:04.810657693 +0000 UTC m=+52.564926567 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810750 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs podName:01650d78-a230-4e2d-80bb-817099bc34f2 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:49.310725335 +0000 UTC m=+37.064994189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs") pod "network-metrics-daemon-kwwvg" (UID: "01650d78-a230-4e2d-80bb-817099bc34f2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810789 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:04.810770596 +0000 UTC m=+52.565039510 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.810983 4969 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: E1004 08:16:48.811075 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:04.811049442 +0000 UTC m=+52.565318336 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.817558 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.838619 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.842192 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-955nm\" (UniqueName: \"kubernetes.io/projected/01650d78-a230-4e2d-80bb-817099bc34f2-kube-api-access-955nm\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.857107 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.876499 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.880220 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.880295 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.880316 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.880690 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.880966 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.895165 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.912279 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.927280 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.948054 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:46Z\\\",\\\"message\\\":\\\" tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z]\\\\nI1004 08:16:46.256960 6427 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 08:16:46.257019 6427 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_router_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:169.254.0.2:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.969986 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.982788 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.982857 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.982868 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.982903 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.982916 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:48Z","lastTransitionTime":"2025-10-04T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:48 crc kubenswrapper[4969]: I1004 08:16:48.993510 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:48Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.007209 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.021345 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.043680 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.054865 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.054931 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.055182 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.055277 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.057854 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.070536 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.082866 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.085300 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.085356 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.085374 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.085397 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.085445 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.101739 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.122976 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.174601 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.174672 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.174691 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.174716 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.174735 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.193416 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.198822 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.198866 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.198882 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.198905 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.198922 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.213183 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.217731 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.217779 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.217794 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.217815 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.217829 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.238730 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.243961 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.244023 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.244047 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.244077 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.244098 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.263486 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.268229 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.268278 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.268292 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.268313 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.268325 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.287976 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:49Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.288165 4969 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.290260 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.290319 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.290340 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.290363 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.290382 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.316847 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.317072 4969 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:49 crc kubenswrapper[4969]: E1004 08:16:49.317170 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs podName:01650d78-a230-4e2d-80bb-817099bc34f2 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:50.3171452 +0000 UTC m=+38.071414054 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs") pod "network-metrics-daemon-kwwvg" (UID: "01650d78-a230-4e2d-80bb-817099bc34f2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.393731 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.393800 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.393821 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.393848 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.393870 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.497492 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.497826 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.497844 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.497872 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.497891 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.601311 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.601379 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.601398 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.601450 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.601471 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.703652 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.703707 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.703718 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.703737 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.703748 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.807113 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.807161 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.807178 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.807200 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.807216 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.909734 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.909795 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.909814 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.909839 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:49 crc kubenswrapper[4969]: I1004 08:16:49.909859 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:49Z","lastTransitionTime":"2025-10-04T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.013623 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.013694 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.013717 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.013746 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.013769 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.054771 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.054771 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:50 crc kubenswrapper[4969]: E1004 08:16:50.054956 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:50 crc kubenswrapper[4969]: E1004 08:16:50.055084 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.116885 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.116944 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.116974 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.117049 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.117072 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.219882 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.219951 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.219969 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.219992 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.220010 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.323649 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.323775 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.323807 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.323839 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.323862 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.330192 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:50 crc kubenswrapper[4969]: E1004 08:16:50.330378 4969 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:50 crc kubenswrapper[4969]: E1004 08:16:50.330509 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs podName:01650d78-a230-4e2d-80bb-817099bc34f2 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:52.330481632 +0000 UTC m=+40.084750486 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs") pod "network-metrics-daemon-kwwvg" (UID: "01650d78-a230-4e2d-80bb-817099bc34f2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.427119 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.427175 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.427192 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.427216 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.427233 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.529851 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.529916 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.529938 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.529968 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.529989 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.581738 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.594087 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.606734 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.618000 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.630814 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.632366 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.632401 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.632410 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.632439 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.632448 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.647353 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.663125 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.680622 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:46Z\\\",\\\"message\\\":\\\" tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z]\\\\nI1004 08:16:46.256960 6427 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 08:16:46.257019 6427 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_router_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:169.254.0.2:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.692098 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.709346 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.722879 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.734835 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.734888 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.734904 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.734927 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.734944 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.738511 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.759807 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.784246 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.800905 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.818600 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.837297 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.837354 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.837364 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.837379 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.837390 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.840008 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:50Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.940231 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.940271 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.940281 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.940294 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:50 crc kubenswrapper[4969]: I1004 08:16:50.940303 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:50Z","lastTransitionTime":"2025-10-04T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.043779 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.043866 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.043882 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.043907 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.043923 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.054496 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.054558 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:51 crc kubenswrapper[4969]: E1004 08:16:51.054659 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:51 crc kubenswrapper[4969]: E1004 08:16:51.054734 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.146630 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.146668 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.146677 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.146708 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.146717 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.249545 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.249595 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.249610 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.249629 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.249640 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.352165 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.352233 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.352247 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.352267 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.352279 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.455594 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.455646 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.455664 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.455688 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.455704 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.558791 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.558855 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.558872 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.558898 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.558916 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.661791 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.661832 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.661844 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.661861 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.661875 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.764548 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.764630 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.764652 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.764679 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.764709 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.867344 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.867378 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.867386 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.867399 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.867408 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.970090 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.970198 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.970235 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.970270 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:51 crc kubenswrapper[4969]: I1004 08:16:51.970292 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:51Z","lastTransitionTime":"2025-10-04T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.054808 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.054872 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:52 crc kubenswrapper[4969]: E1004 08:16:52.055011 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:52 crc kubenswrapper[4969]: E1004 08:16:52.055120 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.073542 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.073574 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.073584 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.073599 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.073610 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.175986 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.176025 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.176033 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.176047 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.176056 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.278252 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.278298 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.278308 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.278326 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.278339 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.354404 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:52 crc kubenswrapper[4969]: E1004 08:16:52.354708 4969 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:52 crc kubenswrapper[4969]: E1004 08:16:52.354803 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs podName:01650d78-a230-4e2d-80bb-817099bc34f2 nodeName:}" failed. No retries permitted until 2025-10-04 08:16:56.354775939 +0000 UTC m=+44.109044793 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs") pod "network-metrics-daemon-kwwvg" (UID: "01650d78-a230-4e2d-80bb-817099bc34f2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.381748 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.381808 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.381824 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.381847 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.381866 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.484920 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.484971 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.484985 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.485003 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.485014 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.587747 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.587824 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.587842 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.587865 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.587882 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.690569 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.690606 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.690614 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.690630 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.690638 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.793039 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.793082 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.793093 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.793110 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.793122 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.896031 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.896081 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.896099 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.896121 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.896137 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.999333 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.999378 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.999389 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.999407 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:52 crc kubenswrapper[4969]: I1004 08:16:52.999447 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:52Z","lastTransitionTime":"2025-10-04T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.054814 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.054901 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:53 crc kubenswrapper[4969]: E1004 08:16:53.054999 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:53 crc kubenswrapper[4969]: E1004 08:16:53.055088 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.073864 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.092196 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.102606 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.102657 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.102668 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.102685 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.102698 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:53Z","lastTransitionTime":"2025-10-04T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.111345 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.131296 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.143982 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.161799 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.175997 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.188858 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.204689 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.204729 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.204741 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.204760 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.204771 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:53Z","lastTransitionTime":"2025-10-04T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.212775 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:46Z\\\",\\\"message\\\":\\\" tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z]\\\\nI1004 08:16:46.256960 6427 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 08:16:46.257019 6427 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_router_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:169.254.0.2:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.244160 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.262587 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.276092 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.290492 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.305532 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.307145 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.307180 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.307190 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.307204 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.307212 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:53Z","lastTransitionTime":"2025-10-04T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.317588 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.328709 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:53Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.410502 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.410562 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.410579 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.410600 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.410618 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:53Z","lastTransitionTime":"2025-10-04T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.513859 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.513911 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.513931 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.513958 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.513982 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:53Z","lastTransitionTime":"2025-10-04T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.616367 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.616477 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.616502 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.616536 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.616559 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:53Z","lastTransitionTime":"2025-10-04T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.718342 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.718384 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.718392 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.718409 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.718439 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:53Z","lastTransitionTime":"2025-10-04T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.820514 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.820593 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.820606 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.820624 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.820638 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:53Z","lastTransitionTime":"2025-10-04T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.923183 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.923225 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.923233 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.923249 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:53 crc kubenswrapper[4969]: I1004 08:16:53.923260 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:53Z","lastTransitionTime":"2025-10-04T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.025879 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.025932 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.025950 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.025969 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.025982 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.054731 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.054746 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:54 crc kubenswrapper[4969]: E1004 08:16:54.054862 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:16:54 crc kubenswrapper[4969]: E1004 08:16:54.055027 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.128863 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.128901 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.128910 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.128924 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.128946 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.232897 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.232959 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.232980 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.233031 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.233050 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.336693 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.336773 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.336795 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.336824 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.336846 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.440345 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.440502 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.440523 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.440550 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.440568 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.543293 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.543329 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.543340 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.543354 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.543363 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.646280 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.646338 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.646355 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.646380 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.646399 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.749892 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.749996 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.750014 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.750038 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.750068 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.853120 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.853190 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.853207 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.853233 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.853250 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.956715 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.956801 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.956826 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.956857 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:54 crc kubenswrapper[4969]: I1004 08:16:54.956881 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:54Z","lastTransitionTime":"2025-10-04T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.055228 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.055223 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:55 crc kubenswrapper[4969]: E1004 08:16:55.055489 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:55 crc kubenswrapper[4969]: E1004 08:16:55.055687 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.059549 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.059635 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.059654 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.059680 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.059697 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.162691 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.162800 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.162824 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.162855 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.162879 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.265607 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.265669 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.265686 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.265709 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.265726 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.369338 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.369416 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.369457 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.369481 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.369496 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.473252 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.473314 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.473328 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.473348 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.473363 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.577186 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.577259 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.577278 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.577307 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.577332 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.681232 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.681303 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.681321 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.681349 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.681365 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.785099 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.785164 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.785180 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.785201 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.785217 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.888281 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.888371 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.888398 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.888470 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.888491 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.992100 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.992187 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.992204 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.992235 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:55 crc kubenswrapper[4969]: I1004 08:16:55.992253 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:55Z","lastTransitionTime":"2025-10-04T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.054299 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.054299 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:56 crc kubenswrapper[4969]: E1004 08:16:56.054617 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:16:56 crc kubenswrapper[4969]: E1004 08:16:56.054769 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.096213 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.096290 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.096315 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.096342 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.096364 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:56Z","lastTransitionTime":"2025-10-04T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.199800 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.199861 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.199880 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.199908 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.199926 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:56Z","lastTransitionTime":"2025-10-04T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.302583 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.302634 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.302646 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.302664 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.302677 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:56Z","lastTransitionTime":"2025-10-04T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.397735 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:56 crc kubenswrapper[4969]: E1004 08:16:56.397917 4969 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:56 crc kubenswrapper[4969]: E1004 08:16:56.398025 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs podName:01650d78-a230-4e2d-80bb-817099bc34f2 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:04.397999503 +0000 UTC m=+52.152268357 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs") pod "network-metrics-daemon-kwwvg" (UID: "01650d78-a230-4e2d-80bb-817099bc34f2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.405460 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.405504 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.405523 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.405546 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.405563 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:56Z","lastTransitionTime":"2025-10-04T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.508601 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.508675 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.508699 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.508728 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.508751 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:56Z","lastTransitionTime":"2025-10-04T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.612256 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.612315 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.612336 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.612360 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.612377 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:56Z","lastTransitionTime":"2025-10-04T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.715125 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.715189 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.715207 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.715231 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.715251 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:56Z","lastTransitionTime":"2025-10-04T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.817899 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.817995 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.818014 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.818037 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.818054 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:56Z","lastTransitionTime":"2025-10-04T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.921129 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.921182 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.921200 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.921223 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:56 crc kubenswrapper[4969]: I1004 08:16:56.921248 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:56Z","lastTransitionTime":"2025-10-04T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.024341 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.024408 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.024476 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.024514 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.024537 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.054943 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.054996 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:57 crc kubenswrapper[4969]: E1004 08:16:57.055151 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:57 crc kubenswrapper[4969]: E1004 08:16:57.055304 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.127792 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.127864 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.127881 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.127910 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.127928 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.231055 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.231122 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.231144 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.231175 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.231196 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.334842 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.334903 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.334919 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.334941 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.334958 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.437812 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.437889 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.437916 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.437982 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.438000 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.541310 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.541377 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.541395 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.541449 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.541467 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.644238 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.644292 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.644310 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.644333 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.644352 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.747458 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.747528 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.747546 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.747572 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.747590 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.851173 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.851248 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.851271 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.851303 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.851326 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.954415 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.954512 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.954531 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.954558 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:57 crc kubenswrapper[4969]: I1004 08:16:57.954577 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:57Z","lastTransitionTime":"2025-10-04T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.054162 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.054223 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:16:58 crc kubenswrapper[4969]: E1004 08:16:58.054367 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:16:58 crc kubenswrapper[4969]: E1004 08:16:58.054478 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.057653 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.057723 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.057747 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.057776 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.057800 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.161013 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.161456 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.161523 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.161558 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.161593 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.264513 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.264583 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.264604 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.264630 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.264653 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.367192 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.367356 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.367381 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.367404 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.367628 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.470219 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.470705 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.470856 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.470988 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.471122 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.574046 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.574347 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.574593 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.574803 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.574944 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.678685 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.678753 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.678771 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.678797 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.678817 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.782563 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.782644 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.782663 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.782690 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.782707 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.886293 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.886353 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.886378 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.886413 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.886485 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.990173 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.990288 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.990307 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.990333 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:58 crc kubenswrapper[4969]: I1004 08:16:58.990362 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:58Z","lastTransitionTime":"2025-10-04T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.054648 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.054681 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:16:59 crc kubenswrapper[4969]: E1004 08:16:59.054861 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:16:59 crc kubenswrapper[4969]: E1004 08:16:59.055013 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.093247 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.093302 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.093322 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.093343 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.093358 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.196026 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.196069 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.196079 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.196095 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.196108 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.298648 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.298719 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.298736 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.298760 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.298782 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.402119 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.402185 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.402210 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.402243 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.402270 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.505714 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.505784 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.505801 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.505827 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.505846 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.608989 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.609052 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.609074 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.609098 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.609117 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.652140 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.652236 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.652265 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.652298 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.652334 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: E1004 08:16:59.678662 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:59Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.683904 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.683979 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.683998 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.684023 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.684053 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: E1004 08:16:59.709246 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:59Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.714860 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.714928 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.714953 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.714983 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.715047 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: E1004 08:16:59.736776 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:59Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.741964 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.742027 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.742046 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.742070 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.742087 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: E1004 08:16:59.763590 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:59Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.768779 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.768910 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.768937 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.768966 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.768987 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: E1004 08:16:59.789578 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:59Z is after 2025-08-24T17:21:41Z" Oct 04 08:16:59 crc kubenswrapper[4969]: E1004 08:16:59.789827 4969 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.792310 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.792359 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.792380 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.792404 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.792448 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.895588 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.895661 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.895756 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.895782 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.895801 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.998236 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.998304 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.998321 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.998343 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:16:59 crc kubenswrapper[4969]: I1004 08:16:59.998358 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:16:59Z","lastTransitionTime":"2025-10-04T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.054216 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.054341 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:00 crc kubenswrapper[4969]: E1004 08:17:00.054473 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:00 crc kubenswrapper[4969]: E1004 08:17:00.054585 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.101687 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.101771 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.101797 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.101830 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.101858 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:00Z","lastTransitionTime":"2025-10-04T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.205073 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.205133 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.205152 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.205176 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.205192 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:00Z","lastTransitionTime":"2025-10-04T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.309380 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.309529 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.309561 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.309589 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.309612 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:00Z","lastTransitionTime":"2025-10-04T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.412498 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.412623 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.412646 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.412673 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.412696 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:00Z","lastTransitionTime":"2025-10-04T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.516235 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.516333 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.516351 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.516376 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.516393 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:00Z","lastTransitionTime":"2025-10-04T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.620241 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.620306 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.620322 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.620345 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.620364 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:00Z","lastTransitionTime":"2025-10-04T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.723096 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.723190 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.723226 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.723262 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.723287 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:00Z","lastTransitionTime":"2025-10-04T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.826404 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.826528 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.826553 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.826586 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.826612 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:00Z","lastTransitionTime":"2025-10-04T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.930360 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.930796 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.930987 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.931127 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:00 crc kubenswrapper[4969]: I1004 08:17:00.931251 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:00Z","lastTransitionTime":"2025-10-04T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.033617 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.033674 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.033691 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.033720 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.033738 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.054955 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.054974 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:01 crc kubenswrapper[4969]: E1004 08:17:01.055192 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:01 crc kubenswrapper[4969]: E1004 08:17:01.055273 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.136287 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.136353 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.136370 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.136395 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.136414 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.239535 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.239594 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.239612 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.239634 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.239653 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.342542 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.342591 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.342601 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.342614 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.342622 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.445879 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.445940 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.445959 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.445986 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.446005 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.549464 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.549521 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.549538 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.549562 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.549580 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.652283 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.652348 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.652365 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.652390 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.652408 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.755148 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.755206 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.755224 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.755250 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.755268 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.857813 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.857883 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.857906 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.857937 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.857958 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.961164 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.961240 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.961264 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.961294 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:01 crc kubenswrapper[4969]: I1004 08:17:01.961320 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:01Z","lastTransitionTime":"2025-10-04T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.054845 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.054920 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:02 crc kubenswrapper[4969]: E1004 08:17:02.055555 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:02 crc kubenswrapper[4969]: E1004 08:17:02.055670 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.056012 4969 scope.go:117] "RemoveContainer" containerID="dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.065334 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.065398 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.065453 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.065485 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.065508 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.169261 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.169685 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.169712 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.169745 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.169774 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.272859 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.272907 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.272927 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.272951 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.272969 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.375525 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.375588 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.375604 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.375625 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.375640 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.416328 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/1.log" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.419778 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.420246 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.436241 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.454317 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.469902 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.481605 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.481657 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.481673 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.481704 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.481726 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.490635 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.512412 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.537032 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:46Z\\\",\\\"message\\\":\\\" tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z]\\\\nI1004 08:16:46.256960 6427 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 08:16:46.257019 6427 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_router_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:169.254.0.2:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.556640 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.581225 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.583789 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.583830 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.583850 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.583866 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.583878 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.595639 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.607531 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.617003 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.625176 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.636998 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.648106 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.659978 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.672910 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:02Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.686781 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.686811 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.686819 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.686833 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.686843 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.789916 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.789963 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.789977 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.789994 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.790008 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.894088 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.894150 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.894167 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.894190 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.894208 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.997602 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.997658 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.997674 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.997698 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:02 crc kubenswrapper[4969]: I1004 08:17:02.997715 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:02Z","lastTransitionTime":"2025-10-04T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.055114 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.055115 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:03 crc kubenswrapper[4969]: E1004 08:17:03.055246 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:03 crc kubenswrapper[4969]: E1004 08:17:03.055305 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.070988 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.083107 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.100109 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.100154 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.100164 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.100178 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.100189 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:03Z","lastTransitionTime":"2025-10-04T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.104481 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:46Z\\\",\\\"message\\\":\\\" tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z]\\\\nI1004 08:16:46.256960 6427 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 08:16:46.257019 6427 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_router_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:169.254.0.2:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.118235 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.138733 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.156201 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.172528 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.186880 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.203054 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.203098 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.203108 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.203125 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.203135 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:03Z","lastTransitionTime":"2025-10-04T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.208753 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.225071 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.245810 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.265831 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.286762 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.305994 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.306051 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.306071 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.306098 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.306117 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:03Z","lastTransitionTime":"2025-10-04T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.308895 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.324966 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.342258 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.409500 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.409556 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.409571 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.409591 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.409603 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:03Z","lastTransitionTime":"2025-10-04T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.425545 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/2.log" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.426327 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/1.log" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.429992 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80" exitCode=1 Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.430057 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.430113 4969 scope.go:117] "RemoveContainer" containerID="dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.430932 4969 scope.go:117] "RemoveContainer" containerID="94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80" Oct 04 08:17:03 crc kubenswrapper[4969]: E1004 08:17:03.431157 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.454457 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.474491 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.486658 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.500516 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.512640 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.512709 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.512726 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.512791 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.512812 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:03Z","lastTransitionTime":"2025-10-04T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.515041 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.529122 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.542371 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.562092 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.582138 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.603451 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.616506 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.616602 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.616632 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.616707 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.616766 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:03Z","lastTransitionTime":"2025-10-04T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.623228 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.642369 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.661258 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.683090 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.701786 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.723937 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.724003 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.724021 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.724047 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.724071 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:03Z","lastTransitionTime":"2025-10-04T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.733835 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca54b26d9d073ff700de0bc7e876c1270974b7b74abcf4799713fc918622165\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:16:46Z\\\",\\\"message\\\":\\\" tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:16:46Z is after 2025-08-24T17:21:41Z]\\\\nI1004 08:16:46.256960 6427 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 08:16:46.257019 6427 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:default/kubernetes]} name:Service_default/kubernetes_TCP_node_router_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:169.254.0.2:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:03Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.827847 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.827956 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.827983 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.828014 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.828036 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:03Z","lastTransitionTime":"2025-10-04T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.931179 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.931262 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.931285 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.931317 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:03 crc kubenswrapper[4969]: I1004 08:17:03.931341 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:03Z","lastTransitionTime":"2025-10-04T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.034872 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.034923 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.034941 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.034964 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.034982 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.054994 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.054996 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.055159 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.055329 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.138729 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.138790 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.138807 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.138835 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.138853 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.242277 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.242342 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.242360 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.242383 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.242404 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.345058 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.345138 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.345162 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.345194 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.345217 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.435865 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/2.log" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.442747 4969 scope.go:117] "RemoveContainer" containerID="94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80" Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.443077 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.446978 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.447084 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.447106 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.447121 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.447133 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.455853 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.469609 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.480527 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.481116 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.481354 4969 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.481408 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs podName:01650d78-a230-4e2d-80bb-817099bc34f2 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:20.481393717 +0000 UTC m=+68.235662531 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs") pod "network-metrics-daemon-kwwvg" (UID: "01650d78-a230-4e2d-80bb-817099bc34f2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.496178 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.508565 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.522168 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.539084 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.550065 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.550137 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.550177 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.550207 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.550229 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.556593 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.587071 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.601715 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.615214 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.634810 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.653861 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.653931 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.653945 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.653990 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.654005 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.656577 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.670652 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.689053 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.704693 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:04Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.757676 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.757759 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.757774 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.757800 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.757814 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.783357 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.783519 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:17:36.78349501 +0000 UTC m=+84.537763814 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.863176 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.863336 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.863382 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.863471 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.863800 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.884195 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.884251 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.884293 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.884334 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.884413 4969 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.884507 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.884527 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.884540 4969 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.884631 4969 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.884684 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.884726 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.884744 4969 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.885249 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:36.88456913 +0000 UTC m=+84.638837984 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.885321 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:36.885302838 +0000 UTC m=+84.639571682 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.885380 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:36.885335439 +0000 UTC m=+84.639604283 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:17:04 crc kubenswrapper[4969]: E1004 08:17:04.885458 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:36.885413771 +0000 UTC m=+84.639682625 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.967141 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.967215 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.967239 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.967268 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:04 crc kubenswrapper[4969]: I1004 08:17:04.967289 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:04Z","lastTransitionTime":"2025-10-04T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.054773 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.054967 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:05 crc kubenswrapper[4969]: E1004 08:17:05.055605 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:05 crc kubenswrapper[4969]: E1004 08:17:05.055358 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.071149 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.071210 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.071226 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.071249 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.071265 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.174388 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.174471 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.174489 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.174514 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.174530 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.277315 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.277371 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.277388 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.277413 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.277474 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.380846 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.380924 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.380951 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.380982 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.381008 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.484099 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.484134 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.484143 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.484157 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.484166 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.586969 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.587023 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.587039 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.587064 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.587081 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.690613 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.690674 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.690692 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.690716 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.690734 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.793695 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.793770 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.793792 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.793819 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.793841 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.896259 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.896326 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.896347 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.896375 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.896397 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.999256 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.999307 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.999319 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.999337 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:05 crc kubenswrapper[4969]: I1004 08:17:05.999350 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:05Z","lastTransitionTime":"2025-10-04T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.054994 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:06 crc kubenswrapper[4969]: E1004 08:17:06.055267 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.054994 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:06 crc kubenswrapper[4969]: E1004 08:17:06.055561 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.102151 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.102228 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.102249 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.102280 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.102302 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:06Z","lastTransitionTime":"2025-10-04T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.205867 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.205950 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.205973 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.206002 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.206022 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:06Z","lastTransitionTime":"2025-10-04T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.309363 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.309487 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.309513 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.309541 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.309563 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:06Z","lastTransitionTime":"2025-10-04T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.412412 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.412487 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.412501 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.412517 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.412529 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:06Z","lastTransitionTime":"2025-10-04T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.515654 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.515730 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.515753 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.515784 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.515806 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:06Z","lastTransitionTime":"2025-10-04T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.618353 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.618485 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.618513 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.618543 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.618569 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:06Z","lastTransitionTime":"2025-10-04T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.722015 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.722088 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.722111 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.722140 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.722162 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:06Z","lastTransitionTime":"2025-10-04T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.825265 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.825370 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.825404 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.825501 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.825526 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:06Z","lastTransitionTime":"2025-10-04T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.929050 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.929108 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.929131 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.929158 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:06 crc kubenswrapper[4969]: I1004 08:17:06.929178 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:06Z","lastTransitionTime":"2025-10-04T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.031832 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.031930 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.031966 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.032001 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.032024 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.054158 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.054165 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:07 crc kubenswrapper[4969]: E1004 08:17:07.054344 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:07 crc kubenswrapper[4969]: E1004 08:17:07.054536 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.135263 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.135320 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.135336 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.135360 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.135400 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.238017 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.238076 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.238096 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.238119 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.238135 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.342390 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.342507 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.342534 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.342568 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.342616 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.445714 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.445791 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.445851 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.445883 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.445905 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.548848 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.548881 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.548889 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.548902 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.548911 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.652502 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.652592 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.652613 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.652642 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.652663 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.756327 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.756485 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.756554 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.756601 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.756628 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.859581 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.859648 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.859667 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.859698 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.859722 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.962775 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.962817 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.962828 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.962843 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:07 crc kubenswrapper[4969]: I1004 08:17:07.962852 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:07Z","lastTransitionTime":"2025-10-04T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.054175 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.054275 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:08 crc kubenswrapper[4969]: E1004 08:17:08.054377 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:08 crc kubenswrapper[4969]: E1004 08:17:08.054640 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.065668 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.065755 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.065788 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.065821 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.065845 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.169458 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.169509 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.169526 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.169553 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.169570 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.272773 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.272811 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.272820 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.272834 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.272844 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.287501 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.302764 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.310378 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.327366 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.341318 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.355837 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.376023 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.376319 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.376350 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.376360 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.376379 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.376390 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.393473 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.412081 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.421470 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.438665 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.451903 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.470190 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.479300 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.479344 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.479359 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.479377 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.479390 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.484982 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.495901 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.506125 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.517908 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.530090 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:08Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.582082 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.582184 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.582198 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.582223 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.582238 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.685625 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.685694 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.685713 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.685739 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.685757 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.788835 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.788930 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.788949 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.788974 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.788993 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.891940 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.892010 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.892035 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.892065 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.892090 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.994517 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.994580 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.994595 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.994618 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:08 crc kubenswrapper[4969]: I1004 08:17:08.994634 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:08Z","lastTransitionTime":"2025-10-04T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.054670 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.054722 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:09 crc kubenswrapper[4969]: E1004 08:17:09.054897 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:09 crc kubenswrapper[4969]: E1004 08:17:09.055051 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.103235 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.103306 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.103328 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.103359 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.103383 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:09Z","lastTransitionTime":"2025-10-04T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.207055 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.207121 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.207138 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.207166 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.207185 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:09Z","lastTransitionTime":"2025-10-04T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.310587 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.310676 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.310700 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.310729 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.310751 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:09Z","lastTransitionTime":"2025-10-04T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.414270 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.414338 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.414358 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.414382 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.414401 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:09Z","lastTransitionTime":"2025-10-04T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.517405 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.517483 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.517499 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.517522 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.517539 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:09Z","lastTransitionTime":"2025-10-04T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.620512 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.620555 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.620566 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.620580 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.620591 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:09Z","lastTransitionTime":"2025-10-04T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.722488 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.722570 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.722586 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.722606 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.722622 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:09Z","lastTransitionTime":"2025-10-04T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.825739 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.825790 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.825808 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.825832 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.825853 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:09Z","lastTransitionTime":"2025-10-04T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.928860 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.928923 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.928939 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.928960 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:09 crc kubenswrapper[4969]: I1004 08:17:09.928979 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:09Z","lastTransitionTime":"2025-10-04T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.031543 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.031623 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.031643 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.031667 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.031686 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.055179 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.055179 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:10 crc kubenswrapper[4969]: E1004 08:17:10.055565 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:10 crc kubenswrapper[4969]: E1004 08:17:10.055374 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.135642 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.135730 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.135754 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.135785 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.135808 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.137765 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.137815 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.137832 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.137856 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.137878 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: E1004 08:17:10.158248 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:10Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.162653 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.162704 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.162718 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.162736 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.162751 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: E1004 08:17:10.175851 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:10Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.179486 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.179520 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.179532 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.179547 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.179559 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: E1004 08:17:10.193239 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:10Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.197588 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.197619 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.197630 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.197648 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.197659 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: E1004 08:17:10.214311 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:10Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.219339 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.219395 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.219414 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.219473 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.219492 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: E1004 08:17:10.238130 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:10Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:10 crc kubenswrapper[4969]: E1004 08:17:10.238357 4969 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.240367 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.240469 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.240490 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.240516 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.240535 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.343415 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.343503 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.343522 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.343547 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.343566 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.445998 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.446070 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.446087 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.446115 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.446133 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.548829 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.548893 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.548912 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.548936 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.548954 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.652672 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.652716 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.652738 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.652764 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.652786 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.756356 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.756462 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.756480 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.756505 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.756521 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.859722 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.859787 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.859806 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.859834 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.859862 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.963473 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.963538 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.963558 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.963583 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:10 crc kubenswrapper[4969]: I1004 08:17:10.963600 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:10Z","lastTransitionTime":"2025-10-04T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.055056 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.055064 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:11 crc kubenswrapper[4969]: E1004 08:17:11.055231 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:11 crc kubenswrapper[4969]: E1004 08:17:11.055304 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.066677 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.066726 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.066743 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.066764 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.066780 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.169655 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.169732 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.169754 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.169784 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.169807 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.272902 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.272944 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.272954 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.272969 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.272978 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.376353 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.376469 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.376496 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.376530 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.376554 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.478794 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.478859 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.478879 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.478903 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.478920 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.581582 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.581649 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.581662 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.581680 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.581691 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.684206 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.684263 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.684279 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.684303 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.684320 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.788087 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.788152 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.788176 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.788207 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.788229 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.891470 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.891540 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.891559 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.891586 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.891604 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.994873 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.994938 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.994956 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.994980 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:11 crc kubenswrapper[4969]: I1004 08:17:11.994998 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:11Z","lastTransitionTime":"2025-10-04T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.055140 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.055171 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:12 crc kubenswrapper[4969]: E1004 08:17:12.055341 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:12 crc kubenswrapper[4969]: E1004 08:17:12.055555 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.098124 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.098200 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.098219 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.098244 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.098261 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:12Z","lastTransitionTime":"2025-10-04T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.201502 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.201578 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.201605 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.201635 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.201657 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:12Z","lastTransitionTime":"2025-10-04T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.304272 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.304468 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.304501 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.304531 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.304553 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:12Z","lastTransitionTime":"2025-10-04T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.407668 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.407763 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.407781 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.407808 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.407825 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:12Z","lastTransitionTime":"2025-10-04T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.510585 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.510678 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.510711 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.510743 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.510766 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:12Z","lastTransitionTime":"2025-10-04T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.613809 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.613851 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.613861 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.613877 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.613891 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:12Z","lastTransitionTime":"2025-10-04T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.716341 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.716382 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.716392 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.716409 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.716439 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:12Z","lastTransitionTime":"2025-10-04T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.819775 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.819829 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.819838 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.819853 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.819863 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:12Z","lastTransitionTime":"2025-10-04T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.921970 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.922018 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.922031 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.922048 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:12 crc kubenswrapper[4969]: I1004 08:17:12.922060 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:12Z","lastTransitionTime":"2025-10-04T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.028023 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.028375 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.028388 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.028406 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.028448 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.055123 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:13 crc kubenswrapper[4969]: E1004 08:17:13.055373 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.055827 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:13 crc kubenswrapper[4969]: E1004 08:17:13.055930 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.079333 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.098029 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8bdcc1b-fe5e-4eac-8777-3e1decad94a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.112058 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.127022 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.130813 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.130858 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.130885 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.130905 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.130914 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.144011 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.157755 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.169510 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.180645 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.190892 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.217087 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.233591 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.233657 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.233680 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.233714 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.233737 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.234668 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.251752 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.264294 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.275806 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.292990 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.306753 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.319031 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:13Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.336900 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.336972 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.336994 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.337022 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.337045 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.440323 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.440381 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.440398 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.440445 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.440463 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.542851 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.542927 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.542950 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.542978 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.543000 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.646243 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.646602 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.646698 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.646763 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.646819 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.749001 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.749258 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.749325 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.749397 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.749491 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.852791 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.852849 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.852868 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.852898 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.852919 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.955505 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.955548 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.955558 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.955576 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:13 crc kubenswrapper[4969]: I1004 08:17:13.955590 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:13Z","lastTransitionTime":"2025-10-04T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.054379 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.054445 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:14 crc kubenswrapper[4969]: E1004 08:17:14.054519 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:14 crc kubenswrapper[4969]: E1004 08:17:14.054744 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.058839 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.059572 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.059732 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.059953 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.060093 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.162583 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.162638 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.162650 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.162669 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.162682 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.265196 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.265242 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.265278 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.265295 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.265307 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.368152 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.368202 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.368218 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.368240 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.368257 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.470171 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.470221 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.470237 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.470272 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.470289 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.574232 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.574512 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.574706 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.574882 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.575016 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.678292 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.678368 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.678390 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.678456 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.678482 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.781152 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.781227 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.781248 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.781274 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.781293 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.884116 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.884557 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.884740 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.884924 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.885366 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.988511 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.988577 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.988596 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.988622 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:14 crc kubenswrapper[4969]: I1004 08:17:14.988641 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:14Z","lastTransitionTime":"2025-10-04T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.054707 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.054835 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:15 crc kubenswrapper[4969]: E1004 08:17:15.054899 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:15 crc kubenswrapper[4969]: E1004 08:17:15.055015 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.091725 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.091785 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.091803 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.091830 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.091848 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:15Z","lastTransitionTime":"2025-10-04T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.194883 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.194943 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.194960 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.194986 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.195004 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:15Z","lastTransitionTime":"2025-10-04T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.297803 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.297855 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.297866 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.297884 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.297897 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:15Z","lastTransitionTime":"2025-10-04T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.400945 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.401030 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.401054 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.401082 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.401100 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:15Z","lastTransitionTime":"2025-10-04T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.503619 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.503688 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.503707 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.503732 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.503750 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:15Z","lastTransitionTime":"2025-10-04T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.606790 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.606851 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.606875 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.606903 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.606926 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:15Z","lastTransitionTime":"2025-10-04T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.709649 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.709713 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.709730 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.709754 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.709771 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:15Z","lastTransitionTime":"2025-10-04T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.813353 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.813472 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.813499 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.813537 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.813560 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:15Z","lastTransitionTime":"2025-10-04T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.916413 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.916520 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.916538 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.916565 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:15 crc kubenswrapper[4969]: I1004 08:17:15.916584 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:15Z","lastTransitionTime":"2025-10-04T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.023638 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.023702 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.023719 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.023739 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.023750 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.054524 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:16 crc kubenswrapper[4969]: E1004 08:17:16.055179 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.055845 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:16 crc kubenswrapper[4969]: E1004 08:17:16.056066 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.127058 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.127102 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.127120 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.127147 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.127168 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.230521 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.230583 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.230606 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.230634 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.230657 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.333100 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.333147 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.333159 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.333175 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.333186 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.435633 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.435682 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.435693 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.435710 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.435719 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.538457 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.538496 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.538505 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.538520 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.538529 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.641632 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.641723 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.641746 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.641770 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.641816 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.745890 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.745976 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.745990 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.746011 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.746050 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.848688 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.848720 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.848728 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.848741 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.848750 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.951140 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.951195 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.951206 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.951225 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:16 crc kubenswrapper[4969]: I1004 08:17:16.951237 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:16Z","lastTransitionTime":"2025-10-04T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.054043 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.054674 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.054752 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.054786 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.054808 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.054695 4969 scope.go:117] "RemoveContainer" containerID="94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.054150 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.054150 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:17 crc kubenswrapper[4969]: E1004 08:17:17.055021 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" Oct 04 08:17:17 crc kubenswrapper[4969]: E1004 08:17:17.055046 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:17 crc kubenswrapper[4969]: E1004 08:17:17.055115 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.157456 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.157501 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.157513 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.157532 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.157546 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.260732 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.260780 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.260796 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.260815 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.260827 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.363548 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.363615 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.363634 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.363658 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.363675 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.465740 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.465780 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.465792 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.465807 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.465819 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.568251 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.568291 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.568302 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.568317 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.568328 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.671359 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.671444 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.671455 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.671469 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.671497 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.773497 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.773525 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.773534 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.773547 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.773555 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.876604 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.877100 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.877316 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.877590 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.877844 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.981032 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.981075 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.981085 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.981102 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:17 crc kubenswrapper[4969]: I1004 08:17:17.981114 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:17Z","lastTransitionTime":"2025-10-04T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.054733 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:18 crc kubenswrapper[4969]: E1004 08:17:18.054825 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.054953 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:18 crc kubenswrapper[4969]: E1004 08:17:18.055013 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.083802 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.083866 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.083883 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.083905 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.083923 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:18Z","lastTransitionTime":"2025-10-04T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.187055 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.187116 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.187132 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.187157 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.187176 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:18Z","lastTransitionTime":"2025-10-04T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.289987 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.290041 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.290058 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.290082 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.290099 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:18Z","lastTransitionTime":"2025-10-04T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.392412 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.392491 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.392508 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.392532 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.392547 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:18Z","lastTransitionTime":"2025-10-04T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.495684 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.495735 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.495745 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.495759 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.495770 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:18Z","lastTransitionTime":"2025-10-04T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.598532 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.598646 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.598669 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.598696 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.598713 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:18Z","lastTransitionTime":"2025-10-04T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.701041 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.701410 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.701660 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.701888 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.702087 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:18Z","lastTransitionTime":"2025-10-04T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.804579 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.804623 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.804634 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.804654 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.804667 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:18Z","lastTransitionTime":"2025-10-04T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.906934 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.906966 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.906974 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.906990 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:18 crc kubenswrapper[4969]: I1004 08:17:18.906998 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:18Z","lastTransitionTime":"2025-10-04T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.009259 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.009300 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.009310 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.009325 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.009336 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.066979 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:19 crc kubenswrapper[4969]: E1004 08:17:19.067106 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.067159 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:19 crc kubenswrapper[4969]: E1004 08:17:19.067247 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.111212 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.111263 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.111279 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.111299 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.111315 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.213816 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.213853 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.213862 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.213876 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.213885 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.315960 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.316038 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.316065 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.316097 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.316121 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.417781 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.417833 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.417845 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.417857 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.417867 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.519890 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.519934 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.519946 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.519964 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.519978 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.623010 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.623036 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.623048 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.623063 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.623075 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.724781 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.724820 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.724832 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.724846 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.724857 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.826748 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.826783 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.826800 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.826819 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.826837 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.929934 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.929966 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.929976 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.930012 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:19 crc kubenswrapper[4969]: I1004 08:17:19.930025 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:19Z","lastTransitionTime":"2025-10-04T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.032266 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.032321 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.032334 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.032348 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.032382 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.054894 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.055037 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.054895 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.055149 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.135881 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.135971 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.135997 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.136030 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.136055 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.239118 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.239162 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.239192 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.239210 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.239221 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.342646 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.342707 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.342723 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.342745 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.342764 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.446319 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.446392 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.446409 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.446455 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.446470 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.482361 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.482657 4969 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.482817 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs podName:01650d78-a230-4e2d-80bb-817099bc34f2 nodeName:}" failed. No retries permitted until 2025-10-04 08:17:52.482784272 +0000 UTC m=+100.237053086 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs") pod "network-metrics-daemon-kwwvg" (UID: "01650d78-a230-4e2d-80bb-817099bc34f2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.511620 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.511658 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.511666 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.511680 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.511690 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.530335 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:20Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.536251 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.536299 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.536335 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.536354 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.536370 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.552987 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:20Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.559834 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.559904 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.559927 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.559958 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.559982 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.580233 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:20Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.584141 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.584199 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.584220 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.584247 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.584270 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.601177 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:20Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.605044 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.605091 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.605103 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.605123 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.605136 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.616738 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:20Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:20 crc kubenswrapper[4969]: E1004 08:17:20.616890 4969 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.618992 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.619067 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.619084 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.619112 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.619134 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.721976 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.722043 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.722073 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.722099 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.722117 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.825281 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.825465 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.825497 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.825529 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.825552 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.927941 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.927994 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.928008 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.928029 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:20 crc kubenswrapper[4969]: I1004 08:17:20.928043 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:20Z","lastTransitionTime":"2025-10-04T08:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.031204 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.031253 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.031263 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.031284 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.031300 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.054115 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.054160 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:21 crc kubenswrapper[4969]: E1004 08:17:21.054244 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:21 crc kubenswrapper[4969]: E1004 08:17:21.054351 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.133955 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.133994 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.134008 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.134028 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.134040 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.236785 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.236817 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.236826 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.236838 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.236846 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.338876 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.338951 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.338972 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.338998 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.339016 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.441868 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.441932 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.441954 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.441983 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.442003 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.544413 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.544472 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.544483 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.544498 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.544509 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.647690 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.647734 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.647742 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.647755 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.647767 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.749894 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.749948 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.749959 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.749972 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.749980 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.852660 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.852745 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.852763 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.852787 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.852808 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.955003 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.955082 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.955106 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.955137 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:21 crc kubenswrapper[4969]: I1004 08:17:21.955167 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:21Z","lastTransitionTime":"2025-10-04T08:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.054515 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.054649 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:22 crc kubenswrapper[4969]: E1004 08:17:22.054818 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:22 crc kubenswrapper[4969]: E1004 08:17:22.054988 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.057164 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.057200 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.057210 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.057225 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.057237 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.159551 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.159620 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.159642 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.159672 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.159695 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.261769 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.261804 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.261815 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.261831 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.261842 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.364064 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.364109 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.364124 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.364145 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.364161 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.466972 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.467025 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.467036 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.467052 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.467062 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.574601 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.574651 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.574667 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.574690 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.574708 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.676943 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.676988 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.677005 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.677029 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.677045 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.781232 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.781333 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.781358 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.781407 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.781467 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.883957 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.884000 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.884016 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.884037 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.884054 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.986528 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.986570 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.986582 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.986595 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:22 crc kubenswrapper[4969]: I1004 08:17:22.986605 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:22Z","lastTransitionTime":"2025-10-04T08:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.054860 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:23 crc kubenswrapper[4969]: E1004 08:17:23.055056 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.054874 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:23 crc kubenswrapper[4969]: E1004 08:17:23.055205 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.071238 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.087885 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8bdcc1b-fe5e-4eac-8777-3e1decad94a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.088388 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.088408 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.088434 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.088453 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.088464 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:23Z","lastTransitionTime":"2025-10-04T08:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.102126 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.113665 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.125303 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.143263 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.158351 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.173373 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.185014 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.190071 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.190199 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.190289 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.190376 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.190482 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:23Z","lastTransitionTime":"2025-10-04T08:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.202979 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.214877 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.224461 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.234436 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.248997 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.260631 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.269958 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.281647 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.293614 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.293658 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.293670 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.293687 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.293700 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:23Z","lastTransitionTime":"2025-10-04T08:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.395840 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.395882 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.395894 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.395909 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.395920 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:23Z","lastTransitionTime":"2025-10-04T08:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.498318 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.498399 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.498452 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.498482 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.498503 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:23Z","lastTransitionTime":"2025-10-04T08:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.505691 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/0.log" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.505734 4969 generic.go:334] "Generic (PLEG): container finished" podID="8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6" containerID="fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e" exitCode=1 Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.505785 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mf8dr" event={"ID":"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6","Type":"ContainerDied","Data":"fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.512410 4969 scope.go:117] "RemoveContainer" containerID="fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.543097 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.563979 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.582853 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.598718 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.600661 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.600722 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.600741 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.600765 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.600782 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:23Z","lastTransitionTime":"2025-10-04T08:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.613055 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.632578 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.651145 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:22Z\\\",\\\"message\\\":\\\"2025-10-04T08:16:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611\\\\n2025-10-04T08:16:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611 to /host/opt/cni/bin/\\\\n2025-10-04T08:16:37Z [verbose] multus-daemon started\\\\n2025-10-04T08:16:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T08:17:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.660923 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.670565 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.683056 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.694468 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8bdcc1b-fe5e-4eac-8777-3e1decad94a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.703475 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.703722 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.703880 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.704006 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.704134 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:23Z","lastTransitionTime":"2025-10-04T08:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.708763 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.720396 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.733817 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.746155 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.757556 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.767907 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:23Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.806142 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.806415 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.806527 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.806616 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.806686 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:23Z","lastTransitionTime":"2025-10-04T08:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.909130 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.909167 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.909176 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.909190 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:23 crc kubenswrapper[4969]: I1004 08:17:23.909198 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:23Z","lastTransitionTime":"2025-10-04T08:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.011903 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.012174 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.012267 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.012365 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.012459 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.054940 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.055021 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:24 crc kubenswrapper[4969]: E1004 08:17:24.055125 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:24 crc kubenswrapper[4969]: E1004 08:17:24.055372 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.115121 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.115353 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.115579 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.115779 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.115912 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.217637 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.217680 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.217697 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.217719 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.217735 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.320268 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.320583 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.320677 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.320783 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.320863 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.423240 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.423276 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.423289 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.423306 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.423319 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.510708 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/0.log" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.510774 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mf8dr" event={"ID":"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6","Type":"ContainerStarted","Data":"c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.525029 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.525367 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.525628 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.525854 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.526077 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.529161 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.546043 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8bdcc1b-fe5e-4eac-8777-3e1decad94a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.562817 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.578432 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.593245 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.606040 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.621165 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.629000 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.629069 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.629093 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.629122 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.629144 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.635173 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.649745 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.668891 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.677824 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.687119 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.701975 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.712563 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.723891 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.730869 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.730981 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.731042 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.731110 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.731184 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.736926 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.749118 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:22Z\\\",\\\"message\\\":\\\"2025-10-04T08:16:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611\\\\n2025-10-04T08:16:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611 to /host/opt/cni/bin/\\\\n2025-10-04T08:16:37Z [verbose] multus-daemon started\\\\n2025-10-04T08:16:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T08:17:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:24Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.834276 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.834311 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.834321 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.834335 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.834346 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.937108 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.937164 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.937177 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.937193 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:24 crc kubenswrapper[4969]: I1004 08:17:24.937204 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:24Z","lastTransitionTime":"2025-10-04T08:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.039516 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.039575 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.039593 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.039617 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.039634 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.054947 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:25 crc kubenswrapper[4969]: E1004 08:17:25.055084 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.055106 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:25 crc kubenswrapper[4969]: E1004 08:17:25.055294 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.141778 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.141809 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.141817 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.141829 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.141837 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.243675 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.243703 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.243712 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.243725 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.243734 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.345807 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.345850 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.345858 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.345872 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.345883 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.448297 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.448339 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.448350 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.448365 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.448377 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.550616 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.550663 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.550680 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.550702 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.550718 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.654168 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.654258 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.654281 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.654311 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.654334 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.756106 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.756165 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.756179 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.756194 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.756205 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.858535 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.858558 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.858567 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.858579 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.858607 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.961583 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.961641 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.961661 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.961686 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:25 crc kubenswrapper[4969]: I1004 08:17:25.961702 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:25Z","lastTransitionTime":"2025-10-04T08:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.054659 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.054785 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:26 crc kubenswrapper[4969]: E1004 08:17:26.054880 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:26 crc kubenswrapper[4969]: E1004 08:17:26.055118 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.064762 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.064794 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.064805 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.064819 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.064830 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.070472 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.167774 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.167858 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.167879 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.167909 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.167927 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.271181 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.271248 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.271266 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.271303 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.271321 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.374463 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.374531 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.374581 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.374607 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.374627 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.477231 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.477293 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.477313 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.477338 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.477355 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.579688 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.579759 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.579777 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.579812 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.579830 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.682553 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.682653 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.682677 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.682701 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.682720 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.785208 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.785253 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.785262 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.785279 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.785289 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.887492 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.887588 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.887600 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.887616 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.887627 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.990271 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.990312 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.990323 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.990341 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:26 crc kubenswrapper[4969]: I1004 08:17:26.990355 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:26Z","lastTransitionTime":"2025-10-04T08:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.054590 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.054688 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:27 crc kubenswrapper[4969]: E1004 08:17:27.054815 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:27 crc kubenswrapper[4969]: E1004 08:17:27.054898 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.092790 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.092852 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.092875 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.092901 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.092924 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:27Z","lastTransitionTime":"2025-10-04T08:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.194713 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.194762 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.194778 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.194798 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.194815 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:27Z","lastTransitionTime":"2025-10-04T08:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.297360 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.297464 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.297484 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.297506 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.297523 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:27Z","lastTransitionTime":"2025-10-04T08:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.401131 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.401206 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.401239 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.401267 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.401288 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:27Z","lastTransitionTime":"2025-10-04T08:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.504621 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.504696 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.504718 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.504742 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.504760 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:27Z","lastTransitionTime":"2025-10-04T08:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.607764 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.607823 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.607842 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.607865 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.607883 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:27Z","lastTransitionTime":"2025-10-04T08:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.711121 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.711174 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.711191 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.711214 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.711230 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:27Z","lastTransitionTime":"2025-10-04T08:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.813908 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.813995 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.814021 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.814051 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.814075 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:27Z","lastTransitionTime":"2025-10-04T08:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.917219 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.917590 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.917788 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.917996 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:27 crc kubenswrapper[4969]: I1004 08:17:27.918182 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:27Z","lastTransitionTime":"2025-10-04T08:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.020809 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.020880 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.020898 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.020925 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.020944 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.054752 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.054773 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:28 crc kubenswrapper[4969]: E1004 08:17:28.055050 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:28 crc kubenswrapper[4969]: E1004 08:17:28.055303 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.123879 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.123940 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.123958 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.123984 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.124010 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.227603 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.227657 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.227675 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.227702 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.227720 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.330755 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.330810 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.330826 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.330850 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.330866 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.432712 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.432739 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.432749 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.432762 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.432771 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.535026 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.535094 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.535114 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.535143 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.535162 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.638660 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.638786 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.638812 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.638841 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.638862 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.742487 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.742547 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.742563 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.742587 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.742606 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.846028 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.846080 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.846096 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.846122 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.846139 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.948522 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.948562 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.948573 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.948591 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:28 crc kubenswrapper[4969]: I1004 08:17:28.948602 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:28Z","lastTransitionTime":"2025-10-04T08:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.051162 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.051250 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.051270 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.051294 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.051312 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:29Z","lastTransitionTime":"2025-10-04T08:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.054736 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:29 crc kubenswrapper[4969]: E1004 08:17:29.054924 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.054974 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:29 crc kubenswrapper[4969]: E1004 08:17:29.055145 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.187732 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.187813 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.187836 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.187865 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.187889 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:29Z","lastTransitionTime":"2025-10-04T08:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.291617 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.291682 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.291703 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.291727 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.291754 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:29Z","lastTransitionTime":"2025-10-04T08:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.394204 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.394260 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.394278 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.394303 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.394321 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:29Z","lastTransitionTime":"2025-10-04T08:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.498026 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.498117 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.498139 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.498171 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.498197 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:29Z","lastTransitionTime":"2025-10-04T08:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.601175 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.601232 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.601246 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.601269 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.601347 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:29Z","lastTransitionTime":"2025-10-04T08:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.705113 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.705193 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.705213 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.705244 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.705265 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:29Z","lastTransitionTime":"2025-10-04T08:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.808204 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.808260 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.808274 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.808293 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.808313 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:29Z","lastTransitionTime":"2025-10-04T08:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.910777 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.910835 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.910852 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.910875 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:29 crc kubenswrapper[4969]: I1004 08:17:29.910892 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:29Z","lastTransitionTime":"2025-10-04T08:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.013732 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.013786 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.013804 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.013827 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.013844 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.054809 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.054838 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:30 crc kubenswrapper[4969]: E1004 08:17:30.055044 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:30 crc kubenswrapper[4969]: E1004 08:17:30.055180 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.116999 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.117077 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.117100 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.117134 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.117156 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.220576 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.220637 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.220654 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.220677 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.220694 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.324510 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.324907 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.325067 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.325232 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.325372 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.427683 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.427729 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.427741 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.427757 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.427770 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.529841 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.530183 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.530388 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.530584 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.530710 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.633243 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.633282 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.633291 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.633307 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.633318 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.737530 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.738162 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.738319 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.738488 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.738646 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.841947 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.842268 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.842410 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.842676 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.842798 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.946155 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.946208 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.946226 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.946250 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:30 crc kubenswrapper[4969]: I1004 08:17:30.946269 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:30Z","lastTransitionTime":"2025-10-04T08:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.010373 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.010679 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.010746 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.010815 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.010908 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: E1004 08:17:31.047973 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.057588 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.057741 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:31 crc kubenswrapper[4969]: E1004 08:17:31.058366 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:31 crc kubenswrapper[4969]: E1004 08:17:31.058556 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.058695 4969 scope.go:117] "RemoveContainer" containerID="94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.059493 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.059535 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.059557 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.059587 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.059610 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: E1004 08:17:31.079911 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.089012 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.089069 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.089082 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.089097 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.089108 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: E1004 08:17:31.102371 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.105145 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.105167 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.105175 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.105187 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.105195 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: E1004 08:17:31.118256 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.121987 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.122016 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.122042 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.122055 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.122065 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: E1004 08:17:31.137191 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: E1004 08:17:31.137349 4969 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.144959 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.144998 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.145009 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.145025 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.145036 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.246806 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.246830 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.246838 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.246850 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.246858 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.349472 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.349505 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.349516 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.349530 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.349541 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.453673 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.453755 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.453771 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.453787 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.453798 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.536993 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/2.log" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.540360 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.541829 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.556971 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.557243 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.557360 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.557506 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.557625 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.561603 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.576182 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.589963 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.603525 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:22Z\\\",\\\"message\\\":\\\"2025-10-04T08:16:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611\\\\n2025-10-04T08:16:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611 to /host/opt/cni/bin/\\\\n2025-10-04T08:16:37Z [verbose] multus-daemon started\\\\n2025-10-04T08:16:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T08:17:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.615407 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.626866 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.642220 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.653822 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8bdcc1b-fe5e-4eac-8777-3e1decad94a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.659784 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.659840 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.659854 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.659874 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.659888 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.669569 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.684233 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.696367 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35a7e8a5-4b73-4102-9902-7a7331d2f5c3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a9c80c537af220b0de87725a1be7e5d7370abd4639e79d49afd8226718aaeed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.709211 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.734689 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.747361 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.760232 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.761867 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.762010 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.762094 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.762183 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.762261 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.777296 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.788179 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.797202 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:31Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.864950 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.865036 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.865062 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.865099 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.865127 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.968226 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.968286 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.968302 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.968325 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:31 crc kubenswrapper[4969]: I1004 08:17:31.968343 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:31Z","lastTransitionTime":"2025-10-04T08:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.054216 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.054244 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:32 crc kubenswrapper[4969]: E1004 08:17:32.054674 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:32 crc kubenswrapper[4969]: E1004 08:17:32.054895 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.071383 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.071751 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.071952 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.072100 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.072234 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:32Z","lastTransitionTime":"2025-10-04T08:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.175740 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.175833 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.175851 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.175876 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.175894 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:32Z","lastTransitionTime":"2025-10-04T08:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.279016 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.279069 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.279087 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.279109 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.279126 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:32Z","lastTransitionTime":"2025-10-04T08:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.382516 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.382585 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.382606 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.382634 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.382654 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:32Z","lastTransitionTime":"2025-10-04T08:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.485948 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.486018 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.486040 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.486069 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.486093 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:32Z","lastTransitionTime":"2025-10-04T08:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.546296 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/3.log" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.547144 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/2.log" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.550788 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d" exitCode=1 Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.550828 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.550865 4969 scope.go:117] "RemoveContainer" containerID="94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.552086 4969 scope.go:117] "RemoveContainer" containerID="bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d" Oct 04 08:17:32 crc kubenswrapper[4969]: E1004 08:17:32.552404 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.571299 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.589592 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.589665 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.589692 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.589726 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.589752 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:32Z","lastTransitionTime":"2025-10-04T08:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.594985 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.611608 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.629453 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.642631 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:22Z\\\",\\\"message\\\":\\\"2025-10-04T08:16:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611\\\\n2025-10-04T08:16:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611 to /host/opt/cni/bin/\\\\n2025-10-04T08:16:37Z [verbose] multus-daemon started\\\\n2025-10-04T08:16:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T08:17:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.652349 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.662867 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.677252 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.689941 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8bdcc1b-fe5e-4eac-8777-3e1decad94a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.692821 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.692850 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.692861 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.692875 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.692886 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:32Z","lastTransitionTime":"2025-10-04T08:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.702121 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.712023 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.723672 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35a7e8a5-4b73-4102-9902-7a7331d2f5c3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a9c80c537af220b0de87725a1be7e5d7370abd4639e79d49afd8226718aaeed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.735377 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.745617 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.760061 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.770095 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.788597 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:32Z\\\",\\\"message\\\":\\\"hub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.003712 6990 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.003852 6990 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.004140 6990 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.004481 6990 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:32.004570 6990 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 08:17:32.004600 6990 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:32.004601 6990 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 08:17:32.004643 6990 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 08:17:32.004668 6990 factory.go:656] Stopping watch factory\\\\nI1004 08:17:32.004688 6990 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.796081 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.796146 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.796164 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.796189 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.796208 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:32Z","lastTransitionTime":"2025-10-04T08:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.807947 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:32Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.898614 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.899163 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.899366 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.899638 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:32 crc kubenswrapper[4969]: I1004 08:17:32.899817 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:32Z","lastTransitionTime":"2025-10-04T08:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.003714 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.003781 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.003799 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.003823 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.003843 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.054602 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.054688 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:33 crc kubenswrapper[4969]: E1004 08:17:33.054811 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:33 crc kubenswrapper[4969]: E1004 08:17:33.055013 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.078042 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.102282 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.106769 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.107047 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.107191 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.107324 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.107535 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.124263 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.146481 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.174339 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.208200 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f3273266eba1f086a97286485905a13b385510845d2e2aba0771fd8c495d80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:03Z\\\",\\\"message\\\":\\\"pping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 08:17:03.052856 6629 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.052998 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053293 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053340 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053388 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.053581 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 08:17:03.054015 6629 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:03.054072 6629 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:03.054110 6629 factory.go:656] Stopping watch factory\\\\nI1004 08:17:03.054125 6629 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:03.054139 6629 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 08:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:32Z\\\",\\\"message\\\":\\\"hub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.003712 6990 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.003852 6990 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.004140 6990 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.004481 6990 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:32.004570 6990 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 08:17:32.004600 6990 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:32.004601 6990 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 08:17:32.004643 6990 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 08:17:32.004668 6990 factory.go:656] Stopping watch factory\\\\nI1004 08:17:32.004688 6990 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.211031 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.211085 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.211102 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.211126 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.211144 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.232924 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.252279 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.270157 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.296606 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.313891 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.313949 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.313967 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.313990 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.314006 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.318774 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:22Z\\\",\\\"message\\\":\\\"2025-10-04T08:16:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611\\\\n2025-10-04T08:16:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611 to /host/opt/cni/bin/\\\\n2025-10-04T08:16:37Z [verbose] multus-daemon started\\\\n2025-10-04T08:16:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T08:17:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.338517 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.359519 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.377482 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35a7e8a5-4b73-4102-9902-7a7331d2f5c3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a9c80c537af220b0de87725a1be7e5d7370abd4639e79d49afd8226718aaeed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.399521 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.417691 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.417743 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.417763 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.417790 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.417808 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.420586 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8bdcc1b-fe5e-4eac-8777-3e1decad94a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.447767 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.469878 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.520900 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.521172 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.521183 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.521200 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.521213 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.558348 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/3.log" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.565465 4969 scope.go:117] "RemoveContainer" containerID="bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d" Oct 04 08:17:33 crc kubenswrapper[4969]: E1004 08:17:33.565790 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.585107 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.604554 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.623129 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.625489 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.625619 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.625646 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.625676 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.625706 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.645637 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.664800 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.700927 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:32Z\\\",\\\"message\\\":\\\"hub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.003712 6990 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.003852 6990 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.004140 6990 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.004481 6990 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:32.004570 6990 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 08:17:32.004600 6990 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:32.004601 6990 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 08:17:32.004643 6990 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 08:17:32.004668 6990 factory.go:656] Stopping watch factory\\\\nI1004 08:17:32.004688 6990 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.726124 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.729579 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.729661 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.729689 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.729723 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.729747 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.748849 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.799184 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.821961 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.832484 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.832557 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.832584 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.832614 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.832637 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.846752 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:22Z\\\",\\\"message\\\":\\\"2025-10-04T08:16:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611\\\\n2025-10-04T08:16:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611 to /host/opt/cni/bin/\\\\n2025-10-04T08:16:37Z [verbose] multus-daemon started\\\\n2025-10-04T08:16:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T08:17:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.862390 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.878848 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.901544 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35a7e8a5-4b73-4102-9902-7a7331d2f5c3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a9c80c537af220b0de87725a1be7e5d7370abd4639e79d49afd8226718aaeed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.920172 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.935414 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.935710 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.935864 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.936024 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.936164 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:33Z","lastTransitionTime":"2025-10-04T08:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.940804 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8bdcc1b-fe5e-4eac-8777-3e1decad94a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.961109 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:33 crc kubenswrapper[4969]: I1004 08:17:33.976825 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:33Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.039067 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.039120 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.039136 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.039157 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.039174 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.054751 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.054952 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:34 crc kubenswrapper[4969]: E1004 08:17:34.055280 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:34 crc kubenswrapper[4969]: E1004 08:17:34.055468 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.142987 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.143106 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.143135 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.143167 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.143188 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.246183 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.246701 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.246913 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.247081 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.247259 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.349819 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.349876 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.349894 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.349917 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.349936 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.452993 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.453316 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.453406 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.453534 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.453613 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.556673 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.556729 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.556748 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.556771 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.556792 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.659811 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.661484 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.661662 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.662307 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.666216 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.769181 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.769631 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.769911 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.770111 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.770308 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.873647 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.873710 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.873732 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.873760 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.873782 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.976727 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.976777 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.976794 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.976813 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:34 crc kubenswrapper[4969]: I1004 08:17:34.976830 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:34Z","lastTransitionTime":"2025-10-04T08:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.054215 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.054263 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:35 crc kubenswrapper[4969]: E1004 08:17:35.054393 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:35 crc kubenswrapper[4969]: E1004 08:17:35.054666 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.079660 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.079715 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.079732 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.079754 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.079771 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:35Z","lastTransitionTime":"2025-10-04T08:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.183245 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.183295 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.183312 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.183334 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.183351 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:35Z","lastTransitionTime":"2025-10-04T08:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.286842 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.286918 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.286935 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.286960 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.286977 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:35Z","lastTransitionTime":"2025-10-04T08:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.389974 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.390019 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.390033 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.390051 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.390064 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:35Z","lastTransitionTime":"2025-10-04T08:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.493053 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.493573 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.493760 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.493985 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.494133 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:35Z","lastTransitionTime":"2025-10-04T08:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.597491 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.597560 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.597582 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.597606 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.597624 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:35Z","lastTransitionTime":"2025-10-04T08:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.700897 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.700959 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.700979 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.701005 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.701022 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:35Z","lastTransitionTime":"2025-10-04T08:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.803705 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.803764 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.803804 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.803826 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.803843 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:35Z","lastTransitionTime":"2025-10-04T08:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.906873 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.907216 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.907393 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.907562 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:35 crc kubenswrapper[4969]: I1004 08:17:35.907732 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:35Z","lastTransitionTime":"2025-10-04T08:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.010413 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.010471 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.010483 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.010503 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.010515 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.054193 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.054207 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.054451 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.054563 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.113499 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.113567 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.113589 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.113616 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.113638 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.223934 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.223962 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.223971 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.223983 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.223992 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.326920 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.326969 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.326979 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.326993 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.327004 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.429383 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.429481 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.429501 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.429525 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.429542 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.532919 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.532990 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.533011 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.533036 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.533053 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.635602 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.635679 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.635697 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.635722 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.635740 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.738726 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.738803 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.738821 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.738845 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.738861 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.841882 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.841977 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.842004 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.842039 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.842063 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.862028 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.862343 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.862315083 +0000 UTC m=+148.616583927 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.945459 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.945508 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.945540 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.945562 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.945574 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:36Z","lastTransitionTime":"2025-10-04T08:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.969057 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.969146 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.969270 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969321 4969 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969417 4969 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969491 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.969401062 +0000 UTC m=+148.723669906 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969569 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.969545376 +0000 UTC m=+148.723814240 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 08:17:36 crc kubenswrapper[4969]: I1004 08:17:36.969336 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969613 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969629 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969665 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969685 4969 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969693 4969 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969713 4969 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969798 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.969773421 +0000 UTC m=+148.724042275 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:17:36 crc kubenswrapper[4969]: E1004 08:17:36.969836 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.969818002 +0000 UTC m=+148.724086856 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.048537 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.048610 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.048634 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.048668 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.048690 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.055134 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.055225 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:37 crc kubenswrapper[4969]: E1004 08:17:37.055297 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:37 crc kubenswrapper[4969]: E1004 08:17:37.055387 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.151780 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.151845 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.151862 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.151886 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.151904 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.255964 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.256065 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.256093 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.256125 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.256161 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.359863 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.360045 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.360078 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.360102 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.360129 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.463281 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.463887 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.464073 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.464270 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.464502 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.567175 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.567219 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.567231 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.567247 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.567259 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.670661 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.670731 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.670753 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.670779 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.670799 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.774703 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.774763 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.774778 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.774803 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.774821 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.878300 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.878353 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.878370 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.878393 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.878415 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.982631 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.982691 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.982704 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.982726 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:37 crc kubenswrapper[4969]: I1004 08:17:37.982742 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:37Z","lastTransitionTime":"2025-10-04T08:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.055038 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.055072 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:38 crc kubenswrapper[4969]: E1004 08:17:38.056035 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:38 crc kubenswrapper[4969]: E1004 08:17:38.056202 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.086656 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.086713 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.086725 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.086749 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.086764 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:38Z","lastTransitionTime":"2025-10-04T08:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.190571 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.191096 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.191241 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.191392 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.191769 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:38Z","lastTransitionTime":"2025-10-04T08:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.295049 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.295136 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.295166 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.295197 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.295219 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:38Z","lastTransitionTime":"2025-10-04T08:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.398025 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.398288 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.398405 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.398553 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.398646 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:38Z","lastTransitionTime":"2025-10-04T08:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.502006 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.502071 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.502086 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.502104 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.502118 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:38Z","lastTransitionTime":"2025-10-04T08:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.604543 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.604628 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.604638 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.604655 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.604665 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:38Z","lastTransitionTime":"2025-10-04T08:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.707522 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.707616 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.707640 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.707671 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.707694 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:38Z","lastTransitionTime":"2025-10-04T08:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.811736 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.811827 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.811847 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.811882 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.811922 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:38Z","lastTransitionTime":"2025-10-04T08:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.916706 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.916751 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.916762 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.916779 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:38 crc kubenswrapper[4969]: I1004 08:17:38.916799 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:38Z","lastTransitionTime":"2025-10-04T08:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.018574 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.018616 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.018625 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.018655 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.018666 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.054498 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:39 crc kubenswrapper[4969]: E1004 08:17:39.054647 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.054498 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:39 crc kubenswrapper[4969]: E1004 08:17:39.055032 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.122481 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.122919 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.122947 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.122976 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.122998 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.225151 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.225196 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.225206 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.225220 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.225232 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.328245 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.328292 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.328303 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.328320 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.328331 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.431169 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.431210 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.431218 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.431231 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.431240 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.533658 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.533706 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.533714 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.533728 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.533737 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.636304 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.636342 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.636352 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.636372 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.636385 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.739222 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.739291 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.739308 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.739338 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.739355 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.842741 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.842806 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.842819 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.842838 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.842851 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.945519 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.945583 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.945600 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.945622 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:39 crc kubenswrapper[4969]: I1004 08:17:39.945639 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:39Z","lastTransitionTime":"2025-10-04T08:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.049169 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.049219 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.049236 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.049254 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.049269 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.054682 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.054715 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:40 crc kubenswrapper[4969]: E1004 08:17:40.054858 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:40 crc kubenswrapper[4969]: E1004 08:17:40.055695 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.152371 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.152487 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.152513 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.152546 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.152569 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.255784 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.256075 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.256105 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.256131 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.256148 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.359493 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.359634 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.359659 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.359686 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.359709 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.462869 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.462921 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.462939 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.462963 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.462980 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.566224 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.566270 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.566285 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.566308 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.566328 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.668878 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.668959 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.668981 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.669089 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.669109 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.772473 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.772567 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.772587 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.772611 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.772628 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.875594 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.875671 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.875683 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.875699 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.875711 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.979041 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.979136 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.979150 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.979166 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:40 crc kubenswrapper[4969]: I1004 08:17:40.979177 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:40Z","lastTransitionTime":"2025-10-04T08:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.054621 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.054705 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:41 crc kubenswrapper[4969]: E1004 08:17:41.054817 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:41 crc kubenswrapper[4969]: E1004 08:17:41.054909 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.081397 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.081440 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.081450 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.081462 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.081472 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.183720 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.183767 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.183778 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.183799 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.183812 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.284652 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.284711 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.284729 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.284751 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.284768 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: E1004 08:17:41.300937 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.304824 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.304854 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.304861 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.304874 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.304883 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: E1004 08:17:41.323167 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.328330 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.328373 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.328385 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.328403 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.328416 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: E1004 08:17:41.341850 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.347104 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.347153 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.347187 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.347208 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.347228 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: E1004 08:17:41.362576 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.366987 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.367037 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.367049 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.367071 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.367083 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: E1004 08:17:41.385318 4969 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"529892e6-c212-481c-813c-a422ee3e502d\\\",\\\"systemUUID\\\":\\\"840d234f-97b2-4b81-bc5d-e0ca9b13faeb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:41Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:41 crc kubenswrapper[4969]: E1004 08:17:41.385482 4969 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.387358 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.387392 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.387449 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.387463 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.387474 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.491030 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.491084 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.491095 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.491113 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.491124 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.594735 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.594803 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.594822 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.594848 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.594866 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.699011 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.699079 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.699095 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.699117 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.699137 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.803178 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.803276 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.803295 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.803318 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.803335 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.907368 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.907462 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.907477 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.907498 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:41 crc kubenswrapper[4969]: I1004 08:17:41.907513 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:41Z","lastTransitionTime":"2025-10-04T08:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.010557 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.010619 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.010630 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.010650 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.010684 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.054288 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:42 crc kubenswrapper[4969]: E1004 08:17:42.054494 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.054928 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:42 crc kubenswrapper[4969]: E1004 08:17:42.055321 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.114349 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.114477 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.114498 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.114525 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.114545 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.218099 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.218758 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.218921 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.219130 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.219281 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.323161 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.323240 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.323261 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.323290 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.323311 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.426963 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.427029 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.427050 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.427079 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.427101 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.530949 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.531026 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.531044 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.531073 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.531092 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.635189 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.635247 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.635259 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.635277 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.635290 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.738216 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.738263 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.738276 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.738291 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.738304 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.840685 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.840731 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.840746 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.840762 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.840772 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.943462 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.943500 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.943511 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.943526 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:42 crc kubenswrapper[4969]: I1004 08:17:42.943537 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:42Z","lastTransitionTime":"2025-10-04T08:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.046985 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.047054 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.047075 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.047103 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.047125 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.054581 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:43 crc kubenswrapper[4969]: E1004 08:17:43.054776 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.054893 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:43 crc kubenswrapper[4969]: E1004 08:17:43.054963 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.074179 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:32Z\\\",\\\"message\\\":\\\"hub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.003712 6990 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.003852 6990 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.004140 6990 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 08:17:32.004481 6990 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 08:17:32.004570 6990 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 08:17:32.004600 6990 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 08:17:32.004601 6990 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 08:17:32.004643 6990 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 08:17:32.004668 6990 factory.go:656] Stopping watch factory\\\\nI1004 08:17:32.004688 6990 ovnkube.go:599] Stopped ovnkube\\\\nI1004 08:17:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:17:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcv2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dl8q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.087746 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dbb40576e79c84655db4795f50be5fa449ed912ef6c39675a953195bcbc6c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.100280 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69d1f843-03d6-403f-8ab3-796e2c97af4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ba683914cc14e023d01ae4e17ff8c56efdda71177955d2520a56f65ae5425c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f75xj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bc2w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.113727 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d55ce576e537c9e15517d7f904b9e60fbef2dbba95fb2fb70f648f9719967bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.124138 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ngkmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16ccc3ae-62c6-4aa2-91f0-9fa55fc6e21f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381211db5bdced59866c8ca1b0570ec8a52a03f7515363540b961b530abe57df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wds2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ngkmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.140269 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5db532e6-047f-472c-92d8-ff80bef1bd19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9be47365c0938fc9cda9f63211e6b68b5a1a70f2ab772bfb8194e58d5ab104f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a28721e337560d4161bc34f1459e3ed1c92e9a267ad4aa5bf7d3b4f415fef625\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ec2b899084310b85485a354e269fc9590a5b8b16345319470356fac71edc113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd02e5d350273ced04bda294f97cf35ffb634b47b621f9f06aa515ae68e6a89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc98dcd49e4124019bbecec1dcc3fd9b8baa53247340091303c9c66008e646ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22bc14c0708d13a96f56be7fe521a3bc368751df5e65e4d9d093019f978d2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a77924dbb2e1905656a4be570eb39b9e7310da322a3ec14d30e549e6f94ac8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flkkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2tw2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.149789 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.149832 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.149846 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.149862 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.149872 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.155450 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mf8dr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T08:17:22Z\\\",\\\"message\\\":\\\"2025-10-04T08:16:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611\\\\n2025-10-04T08:16:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5139bb70-3e56-4cfd-9ba5-a409744d8611 to /host/opt/cni/bin/\\\\n2025-10-04T08:16:37Z [verbose] multus-daemon started\\\\n2025-10-04T08:16:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T08:17:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc784\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mf8dr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.166568 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gz7t7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16b3e242-ff6c-4c8b-aeb2-aed31bfc5eff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9b9cd7e21a6f12b16deed41f07dac76402b6a79420d596874185df0292278d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fb8l7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gz7t7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.178754 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01650d78-a230-4e2d-80bb-817099bc34f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-955nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwwvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.193158 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60b1f39a-b46d-4e9b-bb21-c9d0352b57a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22d0c6573f3cccca587b0c0a9c374cbe3adf9ca18e7cd7fcbc894fb506d09a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41bf78fa424099943c8ba52ca70c885684e503578ce8e77f1c0c6ecfa6091e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d0976a3053134e96d7082a82831799538a6fe8f6b947cc1eb2b97e4dc51662\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://084638dfd27afccfa1b74de23cdc2a63e299bf61dc2df809520c7f759a2868cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2b459e75b25bf3c1c692fedbf7525a8ae8f1c68acc4c3e1ca232d0891a3183\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 08:16:26.649272 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 08:16:26.651368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3354833051/tls.crt::/tmp/serving-cert-3354833051/tls.key\\\\\\\"\\\\nI1004 08:16:32.540776 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 08:16:32.548887 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 08:16:32.548918 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 08:16:32.548952 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 08:16:32.548961 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 08:16:32.556307 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 08:16:32.556332 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556337 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 08:16:32.556342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 08:16:32.556346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 08:16:32.556350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 08:16:32.556367 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 08:16:32.556445 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 08:16:32.559209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8718e2ddb08fe2d7074e11d36f69d79b9f6807107cfdc6303798dfe7e5014a27\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4122071b08dcbc31c606a6170325310644deacab294e74138cf2909e48104ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.204512 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8bdcc1b-fe5e-4eac-8777-3e1decad94a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd1adb69bcae4dff751bedbc0f6cca118efff74455f9e2f184d5d6e109bf3f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef25b70fe944085d942212a9ea73d76b48a427004ff0fba4bae8eeec6e68c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9166392c5d06b2a404d377cee4b2da3ef5368e1a1e2814ea9a5744bfaed7aefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa488dc4a2505cac960ea3937ccb2673c92c4ab6240001f33f229eee7262634b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.215477 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.225324 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.234478 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35a7e8a5-4b73-4102-9902-7a7331d2f5c3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a9c80c537af220b0de87725a1be7e5d7370abd4639e79d49afd8226718aaeed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49cd44f3c8d272262edb2305dd38b69845747310e77efe0df56580a3e632ab2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T08:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.244820 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2da26ba5-d78f-40ac-ba9c-af0616e5f181\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://320d98784a6910756089b18e835081194245e98430c1ac03f13921a79c5556a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323f8de366b5721164583326d182d51ddf0728726b1b029caf845cd565089da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a12d6473565bd78dce1b11b1c030bba14bd52b7fe00829bcacf44c192b303b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d671d9b9962d035138ec926f4a5f946771143974d7e6ee1c507efe706aa35fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.252650 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.252698 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.252707 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.252721 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.252730 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.257739 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad39e34e-f40b-46e1-bf57-544dd3cd15ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b4c9a8ab91675349481e386c7ce69aee404b9ee720a3adfdb6cbdfc2201efd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44f78359c8af5386cda586d0490e8ccce4d7c1b0b5d765955732fe0e06264d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpzbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T08:16:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fgf4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.270017 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09eac9fe013b1f7d56a0409f5fde1cdc4312b2ac2e017337a4af76d8832e61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b8766870c5f5ac867352ddc7acb44afde0c90f53e2bfa3f97133dbd0521029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T08:16:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.280307 4969 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T08:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.354989 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.355032 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.355041 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.355054 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.355064 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.457644 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.457723 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.457772 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.457787 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.457799 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.560648 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.560725 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.560748 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.560780 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.560805 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.665354 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.665467 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.665492 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.665525 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.665548 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.768388 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.768474 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.768486 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.768504 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.768519 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.871847 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.871896 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.871913 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.871935 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.871952 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.975644 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.975707 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.975727 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.975754 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:43 crc kubenswrapper[4969]: I1004 08:17:43.975771 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:43Z","lastTransitionTime":"2025-10-04T08:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.054951 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.054991 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:44 crc kubenswrapper[4969]: E1004 08:17:44.055170 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:44 crc kubenswrapper[4969]: E1004 08:17:44.055358 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.078050 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.078113 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.078171 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.078198 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.078219 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:44Z","lastTransitionTime":"2025-10-04T08:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.181855 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.181913 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.181932 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.181954 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.181971 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:44Z","lastTransitionTime":"2025-10-04T08:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.287094 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.287208 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.287228 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.287257 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.287276 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:44Z","lastTransitionTime":"2025-10-04T08:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.391096 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.391165 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.391188 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.391220 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.391243 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:44Z","lastTransitionTime":"2025-10-04T08:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.494216 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.494263 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.494274 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.494290 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.494305 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:44Z","lastTransitionTime":"2025-10-04T08:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.598325 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.598409 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.598460 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.598493 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.598514 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:44Z","lastTransitionTime":"2025-10-04T08:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.701718 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.701807 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.701869 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.701899 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.701921 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:44Z","lastTransitionTime":"2025-10-04T08:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.806352 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.806407 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.806439 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.806459 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.806472 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:44Z","lastTransitionTime":"2025-10-04T08:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.909935 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.910009 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.910031 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.910058 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:44 crc kubenswrapper[4969]: I1004 08:17:44.910077 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:44Z","lastTransitionTime":"2025-10-04T08:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.013008 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.013079 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.013100 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.013124 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.013142 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.054843 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.054903 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:45 crc kubenswrapper[4969]: E1004 08:17:45.055015 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:45 crc kubenswrapper[4969]: E1004 08:17:45.055129 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.117222 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.117278 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.117297 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.117323 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.117341 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.220815 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.220887 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.220904 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.220929 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.220947 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.324921 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.325019 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.325036 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.325062 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.325080 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.428205 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.428260 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.428277 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.428299 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.428316 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.530881 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.530957 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.530984 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.531012 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.531036 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.634298 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.634354 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.634370 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.634391 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.634409 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.737320 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.737466 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.737507 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.737536 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.737573 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.843349 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.843412 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.843486 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.843529 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.843548 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.947009 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.947057 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.947072 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.947094 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:45 crc kubenswrapper[4969]: I1004 08:17:45.947110 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:45Z","lastTransitionTime":"2025-10-04T08:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.050016 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.050055 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.050065 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.050080 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.050089 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.054294 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.054306 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:46 crc kubenswrapper[4969]: E1004 08:17:46.054442 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:46 crc kubenswrapper[4969]: E1004 08:17:46.054501 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.153546 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.153597 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.153609 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.153625 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.153637 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.255817 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.255881 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.255963 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.256049 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.256081 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.359274 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.359347 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.359381 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.359409 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.359478 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.463241 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.463328 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.463349 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.463378 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.463399 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.569281 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.569351 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.569368 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.569395 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.569412 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.673465 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.673532 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.673554 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.673584 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.673606 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.777341 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.777476 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.777496 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.777526 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.777549 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.880409 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.880524 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.880549 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.880576 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.880595 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.983350 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.983406 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.983465 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.983494 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:46 crc kubenswrapper[4969]: I1004 08:17:46.983514 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:46Z","lastTransitionTime":"2025-10-04T08:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.054828 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.055011 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:47 crc kubenswrapper[4969]: E1004 08:17:47.055249 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:47 crc kubenswrapper[4969]: E1004 08:17:47.055380 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.087630 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.087710 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.087727 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.087756 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.087775 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:47Z","lastTransitionTime":"2025-10-04T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.190784 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.190850 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.190864 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.190889 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.190905 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:47Z","lastTransitionTime":"2025-10-04T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.294917 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.294995 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.295012 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.295036 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.295055 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:47Z","lastTransitionTime":"2025-10-04T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.398582 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.398650 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.398670 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.398694 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.398712 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:47Z","lastTransitionTime":"2025-10-04T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.501908 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.501961 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.501977 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.501999 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.502017 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:47Z","lastTransitionTime":"2025-10-04T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.605394 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.605498 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.605522 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.605545 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.605562 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:47Z","lastTransitionTime":"2025-10-04T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.708800 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.708862 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.708880 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.708903 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.708920 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:47Z","lastTransitionTime":"2025-10-04T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.812408 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.812542 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.812559 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.812581 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.812598 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:47Z","lastTransitionTime":"2025-10-04T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.915927 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.915985 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.916001 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.916024 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:47 crc kubenswrapper[4969]: I1004 08:17:47.916040 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:47Z","lastTransitionTime":"2025-10-04T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.019475 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.019541 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.019560 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.019586 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.019605 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.055140 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.055150 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:48 crc kubenswrapper[4969]: E1004 08:17:48.055943 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:48 crc kubenswrapper[4969]: E1004 08:17:48.056136 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.056925 4969 scope.go:117] "RemoveContainer" containerID="bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d" Oct 04 08:17:48 crc kubenswrapper[4969]: E1004 08:17:48.057186 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.122843 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.122914 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.122933 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.122955 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.122972 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.226735 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.226812 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.226830 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.226856 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.226874 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.329757 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.329828 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.329846 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.329872 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.329897 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.433213 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.433333 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.433362 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.433392 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.433409 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.536976 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.537051 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.537074 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.537106 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.537129 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.640410 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.640616 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.640647 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.640674 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.640695 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.744884 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.744948 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.744966 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.744989 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.745007 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.848109 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.848164 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.848175 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.848191 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.848201 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.950826 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.950901 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.950923 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.950950 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:48 crc kubenswrapper[4969]: I1004 08:17:48.950972 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:48Z","lastTransitionTime":"2025-10-04T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.054124 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.054267 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:49 crc kubenswrapper[4969]: E1004 08:17:49.054374 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.054574 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.054619 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.054641 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.054667 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.054689 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: E1004 08:17:49.055313 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.158524 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.158598 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.158617 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.158642 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.158666 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.263240 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.263299 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.263315 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.263338 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.263358 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.365783 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.365823 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.365833 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.365849 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.365860 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.468726 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.468823 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.468860 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.468877 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.468886 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.571894 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.571963 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.571983 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.572027 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.572045 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.674697 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.674814 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.674833 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.674857 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.674920 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.777641 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.777684 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.777694 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.777711 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.777723 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.881740 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.881810 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.881830 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.881861 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.881887 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.985712 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.985781 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.985797 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.985821 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:49 crc kubenswrapper[4969]: I1004 08:17:49.985844 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:49Z","lastTransitionTime":"2025-10-04T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.054225 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.054507 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:50 crc kubenswrapper[4969]: E1004 08:17:50.054674 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:50 crc kubenswrapper[4969]: E1004 08:17:50.054856 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.074985 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.088993 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.089021 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.089028 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.089041 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.089050 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:50Z","lastTransitionTime":"2025-10-04T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.193068 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.193110 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.193120 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.193135 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.193145 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:50Z","lastTransitionTime":"2025-10-04T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.295758 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.295824 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.295844 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.295870 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.295887 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:50Z","lastTransitionTime":"2025-10-04T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.399525 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.399702 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.399733 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.399824 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.399888 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:50Z","lastTransitionTime":"2025-10-04T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.502970 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.503025 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.503043 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.503117 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.503150 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:50Z","lastTransitionTime":"2025-10-04T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.606578 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.606640 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.606663 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.606690 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.606711 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:50Z","lastTransitionTime":"2025-10-04T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.709231 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.709291 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.709307 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.709332 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.709349 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:50Z","lastTransitionTime":"2025-10-04T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.812008 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.812078 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.812094 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.812119 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.812135 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:50Z","lastTransitionTime":"2025-10-04T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.915933 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.916003 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.916018 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.916039 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:50 crc kubenswrapper[4969]: I1004 08:17:50.916057 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:50Z","lastTransitionTime":"2025-10-04T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.019130 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.019193 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.019213 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.019237 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.019254 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:51Z","lastTransitionTime":"2025-10-04T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.054539 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.054579 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:51 crc kubenswrapper[4969]: E1004 08:17:51.054923 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:51 crc kubenswrapper[4969]: E1004 08:17:51.055065 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.122999 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.123057 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.123074 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.123109 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.123127 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:51Z","lastTransitionTime":"2025-10-04T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.226241 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.226285 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.226302 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.226324 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.226340 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:51Z","lastTransitionTime":"2025-10-04T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.329041 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.329111 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.329129 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.329155 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.329173 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:51Z","lastTransitionTime":"2025-10-04T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.431711 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.431753 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.431764 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.431781 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.431792 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:51Z","lastTransitionTime":"2025-10-04T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.534949 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.535042 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.535091 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.535117 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.535134 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:51Z","lastTransitionTime":"2025-10-04T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.612412 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.612523 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.612576 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.612598 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.612613 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:51Z","lastTransitionTime":"2025-10-04T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.649015 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.649342 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.649569 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.649751 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.649945 4969 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T08:17:51Z","lastTransitionTime":"2025-10-04T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.690747 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7"] Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.692257 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.698024 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.698140 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.698410 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.699884 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.735652 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.735879 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.735969 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.736025 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.736153 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-service-ca\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.784412 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podStartSLOduration=78.784386247 podStartE2EDuration="1m18.784386247s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:51.745144794 +0000 UTC m=+99.499413648" watchObservedRunningTime="2025-10-04 08:17:51.784386247 +0000 UTC m=+99.538655071" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.835525 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-mf8dr" podStartSLOduration=78.835500106 podStartE2EDuration="1m18.835500106s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:51.822313257 +0000 UTC m=+99.576582101" watchObservedRunningTime="2025-10-04 08:17:51.835500106 +0000 UTC m=+99.589768950" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.835768 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-gz7t7" podStartSLOduration=78.835759773 podStartE2EDuration="1m18.835759773s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:51.835067326 +0000 UTC m=+99.589336140" watchObservedRunningTime="2025-10-04 08:17:51.835759773 +0000 UTC m=+99.590028627" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.836954 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.837000 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.837030 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.837053 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-service-ca\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.837106 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.837109 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.837152 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.838556 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-service-ca\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.845671 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.860349 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4fa985b-ac82-442d-81b6-2ff71a4e3e73-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-569q7\" (UID: \"a4fa985b-ac82-442d-81b6-2ff71a4e3e73\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.887929 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.8879057289999999 podStartE2EDuration="1.887905729s" podCreationTimestamp="2025-10-04 08:17:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:51.886594517 +0000 UTC m=+99.640863361" watchObservedRunningTime="2025-10-04 08:17:51.887905729 +0000 UTC m=+99.642174553" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.935718 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.935696838 podStartE2EDuration="1m18.935696838s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:51.911151372 +0000 UTC m=+99.665420246" watchObservedRunningTime="2025-10-04 08:17:51.935696838 +0000 UTC m=+99.689965662" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.948333 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-ngkmm" podStartSLOduration=78.948293624 podStartE2EDuration="1m18.948293624s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:51.948078199 +0000 UTC m=+99.702347043" watchObservedRunningTime="2025-10-04 08:17:51.948293624 +0000 UTC m=+99.702562448" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.967674 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2tw2d" podStartSLOduration=78.967656664 podStartE2EDuration="1m18.967656664s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:51.967358176 +0000 UTC m=+99.721627000" watchObservedRunningTime="2025-10-04 08:17:51.967656664 +0000 UTC m=+99.721925478" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.979276 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.979264495 podStartE2EDuration="25.979264495s" podCreationTimestamp="2025-10-04 08:17:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:51.979209354 +0000 UTC m=+99.733478198" watchObservedRunningTime="2025-10-04 08:17:51.979264495 +0000 UTC m=+99.733533309" Oct 04 08:17:51 crc kubenswrapper[4969]: I1004 08:17:51.997554 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.997534499 podStartE2EDuration="1m15.997534499s" podCreationTimestamp="2025-10-04 08:16:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:51.995258513 +0000 UTC m=+99.749527337" watchObservedRunningTime="2025-10-04 08:17:51.997534499 +0000 UTC m=+99.751803333" Oct 04 08:17:52 crc kubenswrapper[4969]: I1004 08:17:52.009921 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.009884549 podStartE2EDuration="44.009884549s" podCreationTimestamp="2025-10-04 08:17:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:52.009760856 +0000 UTC m=+99.764029670" watchObservedRunningTime="2025-10-04 08:17:52.009884549 +0000 UTC m=+99.764153373" Oct 04 08:17:52 crc kubenswrapper[4969]: I1004 08:17:52.019470 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" Oct 04 08:17:52 crc kubenswrapper[4969]: W1004 08:17:52.032403 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4fa985b_ac82_442d_81b6_2ff71a4e3e73.slice/crio-02a077b96c6c37f19c904f358d7e33f4753100f0f85f7fba9deea26af591ee24 WatchSource:0}: Error finding container 02a077b96c6c37f19c904f358d7e33f4753100f0f85f7fba9deea26af591ee24: Status 404 returned error can't find the container with id 02a077b96c6c37f19c904f358d7e33f4753100f0f85f7fba9deea26af591ee24 Oct 04 08:17:52 crc kubenswrapper[4969]: I1004 08:17:52.054300 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:52 crc kubenswrapper[4969]: I1004 08:17:52.054368 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:52 crc kubenswrapper[4969]: E1004 08:17:52.054467 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:52 crc kubenswrapper[4969]: E1004 08:17:52.054553 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:52 crc kubenswrapper[4969]: I1004 08:17:52.095824 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fgf4t" podStartSLOduration=78.095805923 podStartE2EDuration="1m18.095805923s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:52.095123037 +0000 UTC m=+99.849391851" watchObservedRunningTime="2025-10-04 08:17:52.095805923 +0000 UTC m=+99.850074747" Oct 04 08:17:52 crc kubenswrapper[4969]: I1004 08:17:52.545864 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:52 crc kubenswrapper[4969]: E1004 08:17:52.545992 4969 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:17:52 crc kubenswrapper[4969]: E1004 08:17:52.546458 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs podName:01650d78-a230-4e2d-80bb-817099bc34f2 nodeName:}" failed. No retries permitted until 2025-10-04 08:18:56.546395788 +0000 UTC m=+164.300664642 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs") pod "network-metrics-daemon-kwwvg" (UID: "01650d78-a230-4e2d-80bb-817099bc34f2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 08:17:52 crc kubenswrapper[4969]: I1004 08:17:52.637335 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" event={"ID":"a4fa985b-ac82-442d-81b6-2ff71a4e3e73","Type":"ContainerStarted","Data":"920142d55adc98c4756f5873c8de2ed17c9417cd41cfba3b3dda71035068e94e"} Oct 04 08:17:52 crc kubenswrapper[4969]: I1004 08:17:52.637413 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" event={"ID":"a4fa985b-ac82-442d-81b6-2ff71a4e3e73","Type":"ContainerStarted","Data":"02a077b96c6c37f19c904f358d7e33f4753100f0f85f7fba9deea26af591ee24"} Oct 04 08:17:52 crc kubenswrapper[4969]: I1004 08:17:52.653038 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-569q7" podStartSLOduration=79.653008455 podStartE2EDuration="1m19.653008455s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:17:52.650988466 +0000 UTC m=+100.405257320" watchObservedRunningTime="2025-10-04 08:17:52.653008455 +0000 UTC m=+100.407277309" Oct 04 08:17:53 crc kubenswrapper[4969]: I1004 08:17:53.054791 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:53 crc kubenswrapper[4969]: I1004 08:17:53.054830 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:53 crc kubenswrapper[4969]: E1004 08:17:53.055629 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:53 crc kubenswrapper[4969]: E1004 08:17:53.055762 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:54 crc kubenswrapper[4969]: I1004 08:17:54.054372 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:54 crc kubenswrapper[4969]: I1004 08:17:54.054401 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:54 crc kubenswrapper[4969]: E1004 08:17:54.054625 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:54 crc kubenswrapper[4969]: E1004 08:17:54.054678 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:55 crc kubenswrapper[4969]: I1004 08:17:55.059310 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:55 crc kubenswrapper[4969]: E1004 08:17:55.059513 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:55 crc kubenswrapper[4969]: I1004 08:17:55.059796 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:55 crc kubenswrapper[4969]: E1004 08:17:55.059912 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:56 crc kubenswrapper[4969]: I1004 08:17:56.054563 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:56 crc kubenswrapper[4969]: I1004 08:17:56.054633 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:56 crc kubenswrapper[4969]: E1004 08:17:56.054702 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:56 crc kubenswrapper[4969]: E1004 08:17:56.054835 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:57 crc kubenswrapper[4969]: I1004 08:17:57.054703 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:57 crc kubenswrapper[4969]: I1004 08:17:57.054812 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:57 crc kubenswrapper[4969]: E1004 08:17:57.055016 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:17:57 crc kubenswrapper[4969]: E1004 08:17:57.055156 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:58 crc kubenswrapper[4969]: I1004 08:17:58.054720 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:17:58 crc kubenswrapper[4969]: I1004 08:17:58.054720 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:17:58 crc kubenswrapper[4969]: E1004 08:17:58.054880 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:17:58 crc kubenswrapper[4969]: E1004 08:17:58.054993 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:17:59 crc kubenswrapper[4969]: I1004 08:17:59.055183 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:17:59 crc kubenswrapper[4969]: I1004 08:17:59.055298 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:17:59 crc kubenswrapper[4969]: E1004 08:17:59.055475 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:17:59 crc kubenswrapper[4969]: E1004 08:17:59.055808 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:00 crc kubenswrapper[4969]: I1004 08:18:00.054728 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:00 crc kubenswrapper[4969]: I1004 08:18:00.054761 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:00 crc kubenswrapper[4969]: E1004 08:18:00.054927 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:00 crc kubenswrapper[4969]: E1004 08:18:00.055233 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:01 crc kubenswrapper[4969]: I1004 08:18:01.054295 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:01 crc kubenswrapper[4969]: I1004 08:18:01.054579 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:01 crc kubenswrapper[4969]: E1004 08:18:01.055011 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:01 crc kubenswrapper[4969]: E1004 08:18:01.055240 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:02 crc kubenswrapper[4969]: I1004 08:18:02.054356 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:02 crc kubenswrapper[4969]: I1004 08:18:02.054387 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:02 crc kubenswrapper[4969]: E1004 08:18:02.054571 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:02 crc kubenswrapper[4969]: E1004 08:18:02.055068 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:02 crc kubenswrapper[4969]: I1004 08:18:02.055657 4969 scope.go:117] "RemoveContainer" containerID="bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d" Oct 04 08:18:02 crc kubenswrapper[4969]: E1004 08:18:02.055914 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dl8q8_openshift-ovn-kubernetes(9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" Oct 04 08:18:03 crc kubenswrapper[4969]: I1004 08:18:03.054951 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:03 crc kubenswrapper[4969]: E1004 08:18:03.056986 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:03 crc kubenswrapper[4969]: I1004 08:18:03.057272 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:03 crc kubenswrapper[4969]: E1004 08:18:03.057497 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:04 crc kubenswrapper[4969]: I1004 08:18:04.054732 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:04 crc kubenswrapper[4969]: I1004 08:18:04.054739 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:04 crc kubenswrapper[4969]: E1004 08:18:04.054879 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:04 crc kubenswrapper[4969]: E1004 08:18:04.055476 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:05 crc kubenswrapper[4969]: I1004 08:18:05.054774 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:05 crc kubenswrapper[4969]: I1004 08:18:05.054861 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:05 crc kubenswrapper[4969]: E1004 08:18:05.054896 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:05 crc kubenswrapper[4969]: E1004 08:18:05.055116 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:06 crc kubenswrapper[4969]: I1004 08:18:06.054626 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:06 crc kubenswrapper[4969]: I1004 08:18:06.054694 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:06 crc kubenswrapper[4969]: E1004 08:18:06.054786 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:06 crc kubenswrapper[4969]: E1004 08:18:06.055101 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:07 crc kubenswrapper[4969]: I1004 08:18:07.054360 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:07 crc kubenswrapper[4969]: E1004 08:18:07.054630 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:07 crc kubenswrapper[4969]: I1004 08:18:07.054692 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:07 crc kubenswrapper[4969]: E1004 08:18:07.054838 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:08 crc kubenswrapper[4969]: I1004 08:18:08.054724 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:08 crc kubenswrapper[4969]: I1004 08:18:08.054745 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:08 crc kubenswrapper[4969]: E1004 08:18:08.054898 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:08 crc kubenswrapper[4969]: E1004 08:18:08.055052 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:09 crc kubenswrapper[4969]: I1004 08:18:09.054772 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:09 crc kubenswrapper[4969]: I1004 08:18:09.054833 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:09 crc kubenswrapper[4969]: E1004 08:18:09.054994 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:09 crc kubenswrapper[4969]: E1004 08:18:09.055111 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:09 crc kubenswrapper[4969]: I1004 08:18:09.697000 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/1.log" Oct 04 08:18:09 crc kubenswrapper[4969]: I1004 08:18:09.698281 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/0.log" Oct 04 08:18:09 crc kubenswrapper[4969]: I1004 08:18:09.698362 4969 generic.go:334] "Generic (PLEG): container finished" podID="8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6" containerID="c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995" exitCode=1 Oct 04 08:18:09 crc kubenswrapper[4969]: I1004 08:18:09.698406 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mf8dr" event={"ID":"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6","Type":"ContainerDied","Data":"c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995"} Oct 04 08:18:09 crc kubenswrapper[4969]: I1004 08:18:09.698481 4969 scope.go:117] "RemoveContainer" containerID="fcb01cd0bc13d1bcf38fec9c3ea596341469ebfe8e57bfd715f5aa46a55e4b6e" Oct 04 08:18:09 crc kubenswrapper[4969]: I1004 08:18:09.699120 4969 scope.go:117] "RemoveContainer" containerID="c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995" Oct 04 08:18:09 crc kubenswrapper[4969]: E1004 08:18:09.699388 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-mf8dr_openshift-multus(8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6)\"" pod="openshift-multus/multus-mf8dr" podUID="8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6" Oct 04 08:18:10 crc kubenswrapper[4969]: I1004 08:18:10.054380 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:10 crc kubenswrapper[4969]: I1004 08:18:10.054441 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:10 crc kubenswrapper[4969]: E1004 08:18:10.054589 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:10 crc kubenswrapper[4969]: E1004 08:18:10.054664 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:10 crc kubenswrapper[4969]: I1004 08:18:10.704850 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/1.log" Oct 04 08:18:11 crc kubenswrapper[4969]: I1004 08:18:11.054523 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:11 crc kubenswrapper[4969]: E1004 08:18:11.054664 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:11 crc kubenswrapper[4969]: I1004 08:18:11.054708 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:11 crc kubenswrapper[4969]: E1004 08:18:11.054838 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:12 crc kubenswrapper[4969]: I1004 08:18:12.054792 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:12 crc kubenswrapper[4969]: I1004 08:18:12.054845 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:12 crc kubenswrapper[4969]: E1004 08:18:12.054930 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:12 crc kubenswrapper[4969]: E1004 08:18:12.055060 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:13 crc kubenswrapper[4969]: E1004 08:18:13.011796 4969 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 04 08:18:13 crc kubenswrapper[4969]: I1004 08:18:13.055010 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:13 crc kubenswrapper[4969]: I1004 08:18:13.056067 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:13 crc kubenswrapper[4969]: E1004 08:18:13.059658 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:13 crc kubenswrapper[4969]: E1004 08:18:13.059832 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:13 crc kubenswrapper[4969]: E1004 08:18:13.160342 4969 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 08:18:14 crc kubenswrapper[4969]: I1004 08:18:14.055069 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:14 crc kubenswrapper[4969]: E1004 08:18:14.055219 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:14 crc kubenswrapper[4969]: I1004 08:18:14.055066 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:14 crc kubenswrapper[4969]: E1004 08:18:14.055365 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:15 crc kubenswrapper[4969]: I1004 08:18:15.054539 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:15 crc kubenswrapper[4969]: I1004 08:18:15.054583 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:15 crc kubenswrapper[4969]: E1004 08:18:15.054792 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:15 crc kubenswrapper[4969]: E1004 08:18:15.054928 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:16 crc kubenswrapper[4969]: I1004 08:18:16.054743 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:16 crc kubenswrapper[4969]: I1004 08:18:16.054746 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:16 crc kubenswrapper[4969]: E1004 08:18:16.054930 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:16 crc kubenswrapper[4969]: E1004 08:18:16.055042 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:17 crc kubenswrapper[4969]: I1004 08:18:17.054405 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:17 crc kubenswrapper[4969]: I1004 08:18:17.054648 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:17 crc kubenswrapper[4969]: E1004 08:18:17.055299 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:17 crc kubenswrapper[4969]: I1004 08:18:17.055836 4969 scope.go:117] "RemoveContainer" containerID="bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d" Oct 04 08:18:17 crc kubenswrapper[4969]: E1004 08:18:17.056806 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:17 crc kubenswrapper[4969]: I1004 08:18:17.734031 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/3.log" Oct 04 08:18:17 crc kubenswrapper[4969]: I1004 08:18:17.737281 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerStarted","Data":"a1c38ffbc2e6169e3cdd8e7740ea90186281ace71f537797e1f8c07c93acf1c6"} Oct 04 08:18:17 crc kubenswrapper[4969]: I1004 08:18:17.737722 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:18:17 crc kubenswrapper[4969]: I1004 08:18:17.942558 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podStartSLOduration=103.942528479 podStartE2EDuration="1m43.942528479s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:17.772411021 +0000 UTC m=+125.526679875" watchObservedRunningTime="2025-10-04 08:18:17.942528479 +0000 UTC m=+125.696797343" Oct 04 08:18:17 crc kubenswrapper[4969]: I1004 08:18:17.943184 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kwwvg"] Oct 04 08:18:17 crc kubenswrapper[4969]: I1004 08:18:17.943373 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:17 crc kubenswrapper[4969]: E1004 08:18:17.943612 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:18 crc kubenswrapper[4969]: I1004 08:18:18.054798 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:18 crc kubenswrapper[4969]: E1004 08:18:18.054930 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:18 crc kubenswrapper[4969]: E1004 08:18:18.161679 4969 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 08:18:19 crc kubenswrapper[4969]: I1004 08:18:19.054622 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:19 crc kubenswrapper[4969]: E1004 08:18:19.055171 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:19 crc kubenswrapper[4969]: I1004 08:18:19.054620 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:19 crc kubenswrapper[4969]: E1004 08:18:19.055324 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:20 crc kubenswrapper[4969]: I1004 08:18:20.054497 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:20 crc kubenswrapper[4969]: I1004 08:18:20.054574 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:20 crc kubenswrapper[4969]: E1004 08:18:20.054662 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:20 crc kubenswrapper[4969]: E1004 08:18:20.054898 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:21 crc kubenswrapper[4969]: I1004 08:18:21.054540 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:21 crc kubenswrapper[4969]: I1004 08:18:21.054617 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:21 crc kubenswrapper[4969]: E1004 08:18:21.054954 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:21 crc kubenswrapper[4969]: I1004 08:18:21.055127 4969 scope.go:117] "RemoveContainer" containerID="c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995" Oct 04 08:18:21 crc kubenswrapper[4969]: E1004 08:18:21.055132 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:21 crc kubenswrapper[4969]: I1004 08:18:21.755234 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/1.log" Oct 04 08:18:21 crc kubenswrapper[4969]: I1004 08:18:21.755611 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mf8dr" event={"ID":"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6","Type":"ContainerStarted","Data":"e01ce426ae70129a5c459d559cccad2b6c2be1ece70c9c3f21dfd49482f3cfb1"} Oct 04 08:18:22 crc kubenswrapper[4969]: I1004 08:18:22.054912 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:22 crc kubenswrapper[4969]: I1004 08:18:22.054951 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:22 crc kubenswrapper[4969]: E1004 08:18:22.055129 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 08:18:22 crc kubenswrapper[4969]: E1004 08:18:22.055297 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwwvg" podUID="01650d78-a230-4e2d-80bb-817099bc34f2" Oct 04 08:18:23 crc kubenswrapper[4969]: I1004 08:18:23.054998 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:23 crc kubenswrapper[4969]: I1004 08:18:23.055066 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:23 crc kubenswrapper[4969]: E1004 08:18:23.056975 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 08:18:23 crc kubenswrapper[4969]: E1004 08:18:23.057202 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 08:18:24 crc kubenswrapper[4969]: I1004 08:18:24.054995 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:24 crc kubenswrapper[4969]: I1004 08:18:24.055148 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:24 crc kubenswrapper[4969]: I1004 08:18:24.057367 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 04 08:18:24 crc kubenswrapper[4969]: I1004 08:18:24.060731 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 04 08:18:24 crc kubenswrapper[4969]: I1004 08:18:24.061101 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 04 08:18:24 crc kubenswrapper[4969]: I1004 08:18:24.062882 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 04 08:18:25 crc kubenswrapper[4969]: I1004 08:18:25.054908 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:25 crc kubenswrapper[4969]: I1004 08:18:25.054973 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:25 crc kubenswrapper[4969]: I1004 08:18:25.058206 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 04 08:18:25 crc kubenswrapper[4969]: I1004 08:18:25.058351 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.692615 4969 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.738776 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t4b7l"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.739453 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lrtsv"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.739675 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.740036 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.740413 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.740806 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.741842 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-swbhw"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.742271 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vbgw9"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.742487 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.742772 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.743212 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.743964 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.744931 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.744969 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.746542 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.747054 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.747088 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.747584 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.747606 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.750302 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.750473 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.750590 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.750627 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.750801 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.750868 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.750922 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.751143 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.751161 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.751390 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.751582 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.751602 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.751804 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.751890 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.752085 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.752252 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.752581 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.752603 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.752627 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.752867 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.752921 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.752973 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.752881 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.753873 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.754046 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.754103 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.754284 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.754304 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.754461 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.754613 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755034 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755178 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755277 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755398 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755461 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755549 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755663 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755735 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755922 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755966 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.755935 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.756070 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.758237 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.774348 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.774556 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mfqr6"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.774865 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.774931 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.774943 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mfqr6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.780597 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.785341 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.785588 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.785804 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.786272 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-m67tq"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.787680 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.788659 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.788734 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.788744 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.789753 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.789869 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.790524 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.792333 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.792811 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.804096 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.804545 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xtlw2"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.805111 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.805270 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.805511 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.806047 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hbvs6"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.806607 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.806987 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.807117 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.807935 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.819494 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816036 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.807244 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.808281 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.809072 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816326 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816480 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816518 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816589 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816592 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816638 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816693 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816759 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816782 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816850 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816905 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.816937 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.818763 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.818827 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.818890 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.818892 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.818967 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.825521 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0703cb4b-c7c0-4632-9974-4294afec2b08-serving-cert\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.825701 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0703cb4b-c7c0-4632-9974-4294afec2b08-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.825770 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0703cb4b-c7c0-4632-9974-4294afec2b08-config\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.825806 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdpwn\" (UniqueName: \"kubernetes.io/projected/0703cb4b-c7c0-4632-9974-4294afec2b08-kube-api-access-gdpwn\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.825831 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0703cb4b-c7c0-4632-9974-4294afec2b08-service-ca-bundle\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.826830 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-kxtxq"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.826946 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.827617 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mglbz"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.827718 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.828304 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.828539 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.833176 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.833240 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.834995 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.835140 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.835384 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.835688 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.835988 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cbwg8"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.836170 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.836344 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.836462 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.836543 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.836786 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.836964 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.837095 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.837679 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.837757 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.837988 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.838133 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.838224 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.838363 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.838558 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.838851 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.838884 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.839016 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.838856 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.839168 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.839192 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.839319 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.839460 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.840944 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.841107 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.841202 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.845143 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.845958 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rc9xt"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.846110 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.847138 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.847899 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.848485 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.849025 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.849902 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8xngh"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.850283 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.871001 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.871386 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zps22"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.872607 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.872900 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8kcmp"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.878451 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.880797 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.881903 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.882478 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.883668 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.888789 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.889269 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.889860 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.899698 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lrtsv"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.899775 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.901202 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7tnxs"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.901860 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.901918 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.903026 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.903194 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.903940 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.904817 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.904906 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.907147 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.924382 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t4b7l"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.924447 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vbgw9"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.924530 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926244 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/12536a40-76f7-448a-a0f5-0d79f111da6b-audit-policies\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926290 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnjpr\" (UniqueName: \"kubernetes.io/projected/12536a40-76f7-448a-a0f5-0d79f111da6b-kube-api-access-fnjpr\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926314 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5bb19d8d-559a-4d0a-9009-f1eaf935b36b-profile-collector-cert\") pod \"catalog-operator-68c6474976-6nxfz\" (UID: \"5bb19d8d-559a-4d0a-9009-f1eaf935b36b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926331 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb6w4\" (UniqueName: \"kubernetes.io/projected/5bb19d8d-559a-4d0a-9009-f1eaf935b36b-kube-api-access-nb6w4\") pod \"catalog-operator-68c6474976-6nxfz\" (UID: \"5bb19d8d-559a-4d0a-9009-f1eaf935b36b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926345 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f20abc5-82af-483a-a893-3855b6df7fa5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wtczq\" (UID: \"0f20abc5-82af-483a-a893-3855b6df7fa5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926380 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9ffd\" (UniqueName: \"kubernetes.io/projected/1a654328-5c34-4927-949f-0b5818da9b49-kube-api-access-g9ffd\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926394 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/005c5c3d-9ede-4170-9084-b55e868faab5-audit-dir\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926411 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jpw9\" (UniqueName: \"kubernetes.io/projected/d7581c21-cae9-4c2b-8a46-3c08bcd54180-kube-api-access-5jpw9\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926456 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/12536a40-76f7-448a-a0f5-0d79f111da6b-etcd-client\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926472 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-console-config\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926488 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f20abc5-82af-483a-a893-3855b6df7fa5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wtczq\" (UID: \"0f20abc5-82af-483a-a893-3855b6df7fa5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926505 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b48gz\" (UniqueName: \"kubernetes.io/projected/c77eadcf-e03b-4cc7-901c-3ecd159a47bc-kube-api-access-b48gz\") pod \"package-server-manager-789f6589d5-5mp4f\" (UID: \"c77eadcf-e03b-4cc7-901c-3ecd159a47bc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926521 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/12536a40-76f7-448a-a0f5-0d79f111da6b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926544 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7581c21-cae9-4c2b-8a46-3c08bcd54180-serving-cert\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926559 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde-metrics-tls\") pod \"dns-operator-744455d44c-xtlw2\" (UID: \"e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde\") " pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926585 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-serving-cert\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926599 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wd65\" (UniqueName: \"kubernetes.io/projected/005c5c3d-9ede-4170-9084-b55e868faab5-kube-api-access-5wd65\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926614 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12536a40-76f7-448a-a0f5-0d79f111da6b-serving-cert\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926628 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d632f337-1d3d-42c0-b1b4-3232ed684da4-auth-proxy-config\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926644 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d632f337-1d3d-42c0-b1b4-3232ed684da4-machine-approver-tls\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926662 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0703cb4b-c7c0-4632-9974-4294afec2b08-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926679 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d075b9d0-e883-4f71-a7e6-b5397d983a8b-etcd-service-ca\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926695 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12536a40-76f7-448a-a0f5-0d79f111da6b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926713 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx7nc\" (UniqueName: \"kubernetes.io/projected/5f8ffddc-8c22-4008-850c-b2d9a6738f88-kube-api-access-bx7nc\") pod \"cluster-samples-operator-665b6dd947-p2lsw\" (UID: \"5f8ffddc-8c22-4008-850c-b2d9a6738f88\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926729 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0703cb4b-c7c0-4632-9974-4294afec2b08-config\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926749 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/005c5c3d-9ede-4170-9084-b55e868faab5-etcd-client\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926763 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/12536a40-76f7-448a-a0f5-0d79f111da6b-encryption-config\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926779 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpdch\" (UniqueName: \"kubernetes.io/projected/b4901cb1-78b0-4aa8-9e84-22d3c1f6669c-kube-api-access-cpdch\") pod \"downloads-7954f5f757-mfqr6\" (UID: \"b4901cb1-78b0-4aa8-9e84-22d3c1f6669c\") " pod="openshift-console/downloads-7954f5f757-mfqr6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926793 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926810 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d075b9d0-e883-4f71-a7e6-b5397d983a8b-config\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926826 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdpwn\" (UniqueName: \"kubernetes.io/projected/0703cb4b-c7c0-4632-9974-4294afec2b08-kube-api-access-gdpwn\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926841 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/005c5c3d-9ede-4170-9084-b55e868faab5-encryption-config\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926855 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bqtk\" (UniqueName: \"kubernetes.io/projected/846186fd-c540-464d-b85a-a67f0c8870b6-kube-api-access-2bqtk\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926879 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-config\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926908 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f8ffddc-8c22-4008-850c-b2d9a6738f88-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p2lsw\" (UID: \"5f8ffddc-8c22-4008-850c-b2d9a6738f88\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.926935 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f20abc5-82af-483a-a893-3855b6df7fa5-config\") pod \"kube-controller-manager-operator-78b949d7b-wtczq\" (UID: \"0f20abc5-82af-483a-a893-3855b6df7fa5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.927549 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/846186fd-c540-464d-b85a-a67f0c8870b6-serving-cert\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.927731 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c4fb768-95d5-4587-9e36-cc0d4ce79d5b-serving-cert\") pod \"openshift-config-operator-7777fb866f-nhfgd\" (UID: \"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928411 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0703cb4b-c7c0-4632-9974-4294afec2b08-config\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928419 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0703cb4b-c7c0-4632-9974-4294afec2b08-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928498 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-oauth-config\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928524 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppwwb\" (UniqueName: \"kubernetes.io/projected/7aa45ab6-8235-4e99-af75-c60611ae22f6-kube-api-access-ppwwb\") pod \"openshift-apiserver-operator-796bbdcf4f-qwtnd\" (UID: \"7aa45ab6-8235-4e99-af75-c60611ae22f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928540 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84767\" (UniqueName: \"kubernetes.io/projected/e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde-kube-api-access-84767\") pod \"dns-operator-744455d44c-xtlw2\" (UID: \"e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde\") " pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928620 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa45ab6-8235-4e99-af75-c60611ae22f6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qwtnd\" (UID: \"7aa45ab6-8235-4e99-af75-c60611ae22f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928647 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5bb19d8d-559a-4d0a-9009-f1eaf935b36b-srv-cert\") pod \"catalog-operator-68c6474976-6nxfz\" (UID: \"5bb19d8d-559a-4d0a-9009-f1eaf935b36b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928664 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0be6fb94-7ce5-4556-88f7-0938db54d66e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2ksr9\" (UID: \"0be6fb94-7ce5-4556-88f7-0938db54d66e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928699 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0703cb4b-c7c0-4632-9974-4294afec2b08-serving-cert\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.928717 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-trusted-ca-bundle\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.929643 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-config\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.929691 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.929718 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/12536a40-76f7-448a-a0f5-0d79f111da6b-audit-dir\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.929739 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jwzk\" (UniqueName: \"kubernetes.io/projected/0be6fb94-7ce5-4556-88f7-0938db54d66e-kube-api-access-9jwzk\") pod \"machine-config-controller-84d6567774-2ksr9\" (UID: \"0be6fb94-7ce5-4556-88f7-0938db54d66e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.929996 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-service-ca\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930153 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d7sh\" (UniqueName: \"kubernetes.io/projected/d075b9d0-e883-4f71-a7e6-b5397d983a8b-kube-api-access-5d7sh\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930285 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-client-ca\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930392 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-images\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930515 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930524 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0be6fb94-7ce5-4556-88f7-0938db54d66e-proxy-tls\") pod \"machine-config-controller-84d6567774-2ksr9\" (UID: \"0be6fb94-7ce5-4556-88f7-0938db54d66e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930576 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-image-import-ca\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930594 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aa45ab6-8235-4e99-af75-c60611ae22f6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qwtnd\" (UID: \"7aa45ab6-8235-4e99-af75-c60611ae22f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930676 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mkjx\" (UniqueName: \"kubernetes.io/projected/c5ad8d46-2274-448b-a6bf-e2df545679f9-kube-api-access-5mkjx\") pod \"migrator-59844c95c7-xcfzj\" (UID: \"c5ad8d46-2274-448b-a6bf-e2df545679f9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930732 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930768 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d075b9d0-e883-4f71-a7e6-b5397d983a8b-serving-cert\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930787 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t2ml\" (UniqueName: \"kubernetes.io/projected/2c4fb768-95d5-4587-9e36-cc0d4ce79d5b-kube-api-access-4t2ml\") pod \"openshift-config-operator-7777fb866f-nhfgd\" (UID: \"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930819 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/005c5c3d-9ede-4170-9084-b55e868faab5-serving-cert\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930837 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2c4fb768-95d5-4587-9e36-cc0d4ce79d5b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-nhfgd\" (UID: \"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930856 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-config\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930884 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-oauth-serving-cert\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930900 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bvq8\" (UniqueName: \"kubernetes.io/projected/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-kube-api-access-7bvq8\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930919 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-audit\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930933 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-etcd-serving-ca\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930956 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d075b9d0-e883-4f71-a7e6-b5397d983a8b-etcd-ca\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930971 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d075b9d0-e883-4f71-a7e6-b5397d983a8b-etcd-client\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.930987 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfpm5\" (UniqueName: \"kubernetes.io/projected/d632f337-1d3d-42c0-b1b4-3232ed684da4-kube-api-access-mfpm5\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.931003 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c77eadcf-e03b-4cc7-901c-3ecd159a47bc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5mp4f\" (UID: \"c77eadcf-e03b-4cc7-901c-3ecd159a47bc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.931035 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0703cb4b-c7c0-4632-9974-4294afec2b08-service-ca-bundle\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.931051 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-client-ca\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.931070 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d632f337-1d3d-42c0-b1b4-3232ed684da4-config\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.931086 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/005c5c3d-9ede-4170-9084-b55e868faab5-node-pullsecrets\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.931126 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-config\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.931615 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0703cb4b-c7c0-4632-9974-4294afec2b08-service-ca-bundle\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.933753 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mfqr6"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.934072 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.935160 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0703cb4b-c7c0-4632-9974-4294afec2b08-serving-cert\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.937362 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-swbhw"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.937393 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.938123 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hbvs6"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.939959 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mglbz"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.940222 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.942283 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xtlw2"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.946695 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.947785 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.947921 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2bhvf"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.948614 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-tmqwr"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.949383 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2bhvf" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.949394 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.949922 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.950961 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.952346 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m67tq"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.954236 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.954583 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.956604 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.957498 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.961126 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.961711 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.962126 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.963357 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.964762 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.965856 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.967207 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cbwg8"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.969042 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zps22"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.969082 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.971692 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8xngh"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.976208 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rc9xt"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.981893 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.983549 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8kcmp"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.985120 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.986238 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7tnxs"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.987083 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2bhvf"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.988856 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.992041 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.995024 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.996400 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-4qbp6"] Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.997109 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:32 crc kubenswrapper[4969]: I1004 08:18:32.997911 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tmqwr"] Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.003326 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.012110 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-x9p72"] Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.014099 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.020776 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-x9p72"] Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.021988 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.032546 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-audit\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.032574 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-etcd-serving-ca\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.032594 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d075b9d0-e883-4f71-a7e6-b5397d983a8b-etcd-ca\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.032611 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d075b9d0-e883-4f71-a7e6-b5397d983a8b-etcd-client\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.032628 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfpm5\" (UniqueName: \"kubernetes.io/projected/d632f337-1d3d-42c0-b1b4-3232ed684da4-kube-api-access-mfpm5\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.032645 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c77eadcf-e03b-4cc7-901c-3ecd159a47bc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5mp4f\" (UID: \"c77eadcf-e03b-4cc7-901c-3ecd159a47bc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.032662 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-client-ca\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.032676 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d632f337-1d3d-42c0-b1b4-3232ed684da4-config\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033227 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/005c5c3d-9ede-4170-9084-b55e868faab5-node-pullsecrets\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033245 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-config\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033262 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/12536a40-76f7-448a-a0f5-0d79f111da6b-audit-policies\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033277 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnjpr\" (UniqueName: \"kubernetes.io/projected/12536a40-76f7-448a-a0f5-0d79f111da6b-kube-api-access-fnjpr\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033292 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5bb19d8d-559a-4d0a-9009-f1eaf935b36b-profile-collector-cert\") pod \"catalog-operator-68c6474976-6nxfz\" (UID: \"5bb19d8d-559a-4d0a-9009-f1eaf935b36b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033309 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb6w4\" (UniqueName: \"kubernetes.io/projected/5bb19d8d-559a-4d0a-9009-f1eaf935b36b-kube-api-access-nb6w4\") pod \"catalog-operator-68c6474976-6nxfz\" (UID: \"5bb19d8d-559a-4d0a-9009-f1eaf935b36b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033326 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f20abc5-82af-483a-a893-3855b6df7fa5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wtczq\" (UID: \"0f20abc5-82af-483a-a893-3855b6df7fa5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033344 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9ffd\" (UniqueName: \"kubernetes.io/projected/1a654328-5c34-4927-949f-0b5818da9b49-kube-api-access-g9ffd\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033377 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/005c5c3d-9ede-4170-9084-b55e868faab5-audit-dir\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033394 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jpw9\" (UniqueName: \"kubernetes.io/projected/d7581c21-cae9-4c2b-8a46-3c08bcd54180-kube-api-access-5jpw9\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033409 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/12536a40-76f7-448a-a0f5-0d79f111da6b-etcd-client\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033448 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f20abc5-82af-483a-a893-3855b6df7fa5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wtczq\" (UID: \"0f20abc5-82af-483a-a893-3855b6df7fa5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033464 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b48gz\" (UniqueName: \"kubernetes.io/projected/c77eadcf-e03b-4cc7-901c-3ecd159a47bc-kube-api-access-b48gz\") pod \"package-server-manager-789f6589d5-5mp4f\" (UID: \"c77eadcf-e03b-4cc7-901c-3ecd159a47bc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033481 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-console-config\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033495 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/12536a40-76f7-448a-a0f5-0d79f111da6b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033540 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7581c21-cae9-4c2b-8a46-3c08bcd54180-serving-cert\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033555 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde-metrics-tls\") pod \"dns-operator-744455d44c-xtlw2\" (UID: \"e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde\") " pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033577 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-serving-cert\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033609 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wd65\" (UniqueName: \"kubernetes.io/projected/005c5c3d-9ede-4170-9084-b55e868faab5-kube-api-access-5wd65\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033623 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12536a40-76f7-448a-a0f5-0d79f111da6b-serving-cert\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033642 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d632f337-1d3d-42c0-b1b4-3232ed684da4-auth-proxy-config\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033655 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-etcd-serving-ca\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033657 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d632f337-1d3d-42c0-b1b4-3232ed684da4-machine-approver-tls\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033702 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d075b9d0-e883-4f71-a7e6-b5397d983a8b-etcd-service-ca\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033721 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12536a40-76f7-448a-a0f5-0d79f111da6b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033741 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx7nc\" (UniqueName: \"kubernetes.io/projected/5f8ffddc-8c22-4008-850c-b2d9a6738f88-kube-api-access-bx7nc\") pod \"cluster-samples-operator-665b6dd947-p2lsw\" (UID: \"5f8ffddc-8c22-4008-850c-b2d9a6738f88\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033761 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/005c5c3d-9ede-4170-9084-b55e868faab5-etcd-client\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033758 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-audit\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033787 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/005c5c3d-9ede-4170-9084-b55e868faab5-node-pullsecrets\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033777 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/12536a40-76f7-448a-a0f5-0d79f111da6b-encryption-config\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033848 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpdch\" (UniqueName: \"kubernetes.io/projected/b4901cb1-78b0-4aa8-9e84-22d3c1f6669c-kube-api-access-cpdch\") pod \"downloads-7954f5f757-mfqr6\" (UID: \"b4901cb1-78b0-4aa8-9e84-22d3c1f6669c\") " pod="openshift-console/downloads-7954f5f757-mfqr6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033881 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033887 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d632f337-1d3d-42c0-b1b4-3232ed684da4-config\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033921 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/005c5c3d-9ede-4170-9084-b55e868faab5-encryption-config\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033969 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bqtk\" (UniqueName: \"kubernetes.io/projected/846186fd-c540-464d-b85a-a67f0c8870b6-kube-api-access-2bqtk\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033993 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-config\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034012 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d075b9d0-e883-4f71-a7e6-b5397d983a8b-config\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034030 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f8ffddc-8c22-4008-850c-b2d9a6738f88-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p2lsw\" (UID: \"5f8ffddc-8c22-4008-850c-b2d9a6738f88\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034053 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f20abc5-82af-483a-a893-3855b6df7fa5-config\") pod \"kube-controller-manager-operator-78b949d7b-wtczq\" (UID: \"0f20abc5-82af-483a-a893-3855b6df7fa5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034069 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c4fb768-95d5-4587-9e36-cc0d4ce79d5b-serving-cert\") pod \"openshift-config-operator-7777fb866f-nhfgd\" (UID: \"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034085 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/846186fd-c540-464d-b85a-a67f0c8870b6-serving-cert\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034102 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-oauth-config\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034120 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppwwb\" (UniqueName: \"kubernetes.io/projected/7aa45ab6-8235-4e99-af75-c60611ae22f6-kube-api-access-ppwwb\") pod \"openshift-apiserver-operator-796bbdcf4f-qwtnd\" (UID: \"7aa45ab6-8235-4e99-af75-c60611ae22f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034136 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84767\" (UniqueName: \"kubernetes.io/projected/e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde-kube-api-access-84767\") pod \"dns-operator-744455d44c-xtlw2\" (UID: \"e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde\") " pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034153 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-client-ca\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034162 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa45ab6-8235-4e99-af75-c60611ae22f6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qwtnd\" (UID: \"7aa45ab6-8235-4e99-af75-c60611ae22f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034237 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5bb19d8d-559a-4d0a-9009-f1eaf935b36b-srv-cert\") pod \"catalog-operator-68c6474976-6nxfz\" (UID: \"5bb19d8d-559a-4d0a-9009-f1eaf935b36b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034271 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-trusted-ca-bundle\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034298 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-config\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034322 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-config\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034325 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034370 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/12536a40-76f7-448a-a0f5-0d79f111da6b-audit-dir\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034399 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0be6fb94-7ce5-4556-88f7-0938db54d66e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2ksr9\" (UID: \"0be6fb94-7ce5-4556-88f7-0938db54d66e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034460 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jwzk\" (UniqueName: \"kubernetes.io/projected/0be6fb94-7ce5-4556-88f7-0938db54d66e-kube-api-access-9jwzk\") pod \"machine-config-controller-84d6567774-2ksr9\" (UID: \"0be6fb94-7ce5-4556-88f7-0938db54d66e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034488 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-service-ca\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034515 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d7sh\" (UniqueName: \"kubernetes.io/projected/d075b9d0-e883-4f71-a7e6-b5397d983a8b-kube-api-access-5d7sh\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034539 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-client-ca\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034564 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-images\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034606 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0be6fb94-7ce5-4556-88f7-0938db54d66e-proxy-tls\") pod \"machine-config-controller-84d6567774-2ksr9\" (UID: \"0be6fb94-7ce5-4556-88f7-0938db54d66e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034626 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aa45ab6-8235-4e99-af75-c60611ae22f6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qwtnd\" (UID: \"7aa45ab6-8235-4e99-af75-c60611ae22f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034643 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-image-import-ca\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034669 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mkjx\" (UniqueName: \"kubernetes.io/projected/c5ad8d46-2274-448b-a6bf-e2df545679f9-kube-api-access-5mkjx\") pod \"migrator-59844c95c7-xcfzj\" (UID: \"c5ad8d46-2274-448b-a6bf-e2df545679f9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034701 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034719 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d075b9d0-e883-4f71-a7e6-b5397d983a8b-serving-cert\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034735 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t2ml\" (UniqueName: \"kubernetes.io/projected/2c4fb768-95d5-4587-9e36-cc0d4ce79d5b-kube-api-access-4t2ml\") pod \"openshift-config-operator-7777fb866f-nhfgd\" (UID: \"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034755 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/005c5c3d-9ede-4170-9084-b55e868faab5-serving-cert\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034773 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2c4fb768-95d5-4587-9e36-cc0d4ce79d5b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-nhfgd\" (UID: \"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034789 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-config\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034812 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-oauth-serving-cert\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.034829 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bvq8\" (UniqueName: \"kubernetes.io/projected/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-kube-api-access-7bvq8\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.035336 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-trusted-ca-bundle\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.035408 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/12536a40-76f7-448a-a0f5-0d79f111da6b-audit-policies\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.035593 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.035850 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d075b9d0-e883-4f71-a7e6-b5397d983a8b-etcd-client\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.036202 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/005c5c3d-9ede-4170-9084-b55e868faab5-encryption-config\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.036578 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d632f337-1d3d-42c0-b1b4-3232ed684da4-machine-approver-tls\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.036748 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/12536a40-76f7-448a-a0f5-0d79f111da6b-encryption-config\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.036891 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.037221 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d075b9d0-e883-4f71-a7e6-b5397d983a8b-etcd-service-ca\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.037517 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f20abc5-82af-483a-a893-3855b6df7fa5-config\") pod \"kube-controller-manager-operator-78b949d7b-wtczq\" (UID: \"0f20abc5-82af-483a-a893-3855b6df7fa5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.037578 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12536a40-76f7-448a-a0f5-0d79f111da6b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.038126 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-config\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.038248 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-config\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.038307 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/005c5c3d-9ede-4170-9084-b55e868faab5-audit-dir\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.038345 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/12536a40-76f7-448a-a0f5-0d79f111da6b-audit-dir\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.038701 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d075b9d0-e883-4f71-a7e6-b5397d983a8b-config\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.039224 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0be6fb94-7ce5-4556-88f7-0938db54d66e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2ksr9\" (UID: \"0be6fb94-7ce5-4556-88f7-0938db54d66e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.039599 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/12536a40-76f7-448a-a0f5-0d79f111da6b-etcd-client\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.039913 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/005c5c3d-9ede-4170-9084-b55e868faab5-etcd-client\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.039933 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-service-ca\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.039934 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7581c21-cae9-4c2b-8a46-3c08bcd54180-serving-cert\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.040525 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-console-config\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.040579 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde-metrics-tls\") pod \"dns-operator-744455d44c-xtlw2\" (UID: \"e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde\") " pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.033616 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d075b9d0-e883-4f71-a7e6-b5397d983a8b-etcd-ca\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.040747 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-client-ca\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.040995 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/12536a40-76f7-448a-a0f5-0d79f111da6b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.041055 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d632f337-1d3d-42c0-b1b4-3232ed684da4-auth-proxy-config\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.042669 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f8ffddc-8c22-4008-850c-b2d9a6738f88-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p2lsw\" (UID: \"5f8ffddc-8c22-4008-850c-b2d9a6738f88\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.042964 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12536a40-76f7-448a-a0f5-0d79f111da6b-serving-cert\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.043204 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.043373 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/005c5c3d-9ede-4170-9084-b55e868faab5-image-import-ca\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.043962 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aa45ab6-8235-4e99-af75-c60611ae22f6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qwtnd\" (UID: \"7aa45ab6-8235-4e99-af75-c60611ae22f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.044046 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c4fb768-95d5-4587-9e36-cc0d4ce79d5b-serving-cert\") pod \"openshift-config-operator-7777fb866f-nhfgd\" (UID: \"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.044334 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2c4fb768-95d5-4587-9e36-cc0d4ce79d5b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-nhfgd\" (UID: \"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.044912 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa45ab6-8235-4e99-af75-c60611ae22f6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qwtnd\" (UID: \"7aa45ab6-8235-4e99-af75-c60611ae22f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.045617 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-serving-cert\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.045691 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d075b9d0-e883-4f71-a7e6-b5397d983a8b-serving-cert\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.045728 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f20abc5-82af-483a-a893-3855b6df7fa5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wtczq\" (UID: \"0f20abc5-82af-483a-a893-3855b6df7fa5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.046112 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-oauth-config\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.047093 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-config\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.047282 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/846186fd-c540-464d-b85a-a67f0c8870b6-serving-cert\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.047728 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/005c5c3d-9ede-4170-9084-b55e868faab5-serving-cert\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.049138 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-oauth-serving-cert\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.057769 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.057766 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-images\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.061796 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.081603 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.102056 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.121896 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.141743 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.163040 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.181847 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.202157 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.231553 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.243066 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.262443 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.283616 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.302804 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.323757 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.346672 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.363014 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.383327 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.404032 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.422998 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.432078 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0be6fb94-7ce5-4556-88f7-0938db54d66e-proxy-tls\") pod \"machine-config-controller-84d6567774-2ksr9\" (UID: \"0be6fb94-7ce5-4556-88f7-0938db54d66e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.443814 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.483459 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.504476 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.522721 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.542934 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.563654 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.582403 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.602578 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.623380 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.634385 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5bb19d8d-559a-4d0a-9009-f1eaf935b36b-srv-cert\") pod \"catalog-operator-68c6474976-6nxfz\" (UID: \"5bb19d8d-559a-4d0a-9009-f1eaf935b36b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.643762 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.651651 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5bb19d8d-559a-4d0a-9009-f1eaf935b36b-profile-collector-cert\") pod \"catalog-operator-68c6474976-6nxfz\" (UID: \"5bb19d8d-559a-4d0a-9009-f1eaf935b36b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.663971 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.684240 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.703019 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.723367 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.743043 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.763913 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.782650 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.803277 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.843366 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.846742 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.861261 4969 request.go:700] Waited for 1.010471074s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/secrets?fieldSelector=metadata.name%3Dv4-0-config-user-template-login&limit=500&resourceVersion=0 Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.872870 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.895427 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.903293 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.923864 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.942761 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.963113 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 04 08:18:33 crc kubenswrapper[4969]: I1004 08:18:33.984889 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.004066 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.022890 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 04 08:18:34 crc kubenswrapper[4969]: E1004 08:18:34.033203 4969 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 08:18:34 crc kubenswrapper[4969]: E1004 08:18:34.033289 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c77eadcf-e03b-4cc7-901c-3ecd159a47bc-package-server-manager-serving-cert podName:c77eadcf-e03b-4cc7-901c-3ecd159a47bc nodeName:}" failed. No retries permitted until 2025-10-04 08:18:34.533265155 +0000 UTC m=+142.287533989 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/c77eadcf-e03b-4cc7-901c-3ecd159a47bc-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-5mp4f" (UID: "c77eadcf-e03b-4cc7-901c-3ecd159a47bc") : failed to sync secret cache: timed out waiting for the condition Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.042536 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.063259 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.083362 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.104076 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.122926 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.142917 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.164819 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.183409 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.203844 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.224112 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.243029 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.263614 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.290642 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.303651 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.323422 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.342704 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.363514 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.383520 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.403989 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.424184 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.454734 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.462324 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.484511 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.503299 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.555622 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdpwn\" (UniqueName: \"kubernetes.io/projected/0703cb4b-c7c0-4632-9974-4294afec2b08-kube-api-access-gdpwn\") pod \"authentication-operator-69f744f599-vbgw9\" (UID: \"0703cb4b-c7c0-4632-9974-4294afec2b08\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.558451 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c77eadcf-e03b-4cc7-901c-3ecd159a47bc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5mp4f\" (UID: \"c77eadcf-e03b-4cc7-901c-3ecd159a47bc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.563905 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c77eadcf-e03b-4cc7-901c-3ecd159a47bc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5mp4f\" (UID: \"c77eadcf-e03b-4cc7-901c-3ecd159a47bc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.564930 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.583163 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.602400 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.622868 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.643569 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.663268 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.683585 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.704420 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.709508 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.723446 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.743876 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.763580 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.783496 4969 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.803161 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.861705 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfpm5\" (UniqueName: \"kubernetes.io/projected/d632f337-1d3d-42c0-b1b4-3232ed684da4-kube-api-access-mfpm5\") pod \"machine-approver-56656f9798-6b7p5\" (UID: \"d632f337-1d3d-42c0-b1b4-3232ed684da4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.876668 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jpw9\" (UniqueName: \"kubernetes.io/projected/d7581c21-cae9-4c2b-8a46-3c08bcd54180-kube-api-access-5jpw9\") pod \"route-controller-manager-6576b87f9c-rshw5\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.880789 4969 request.go:700] Waited for 1.845259595s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/serviceaccounts/oauth-apiserver-sa/token Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.896589 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bvq8\" (UniqueName: \"kubernetes.io/projected/5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8-kube-api-access-7bvq8\") pod \"machine-api-operator-5694c8668f-swbhw\" (UID: \"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.898451 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnjpr\" (UniqueName: \"kubernetes.io/projected/12536a40-76f7-448a-a0f5-0d79f111da6b-kube-api-access-fnjpr\") pod \"apiserver-7bbb656c7d-xcxzt\" (UID: \"12536a40-76f7-448a-a0f5-0d79f111da6b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.924864 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb6w4\" (UniqueName: \"kubernetes.io/projected/5bb19d8d-559a-4d0a-9009-f1eaf935b36b-kube-api-access-nb6w4\") pod \"catalog-operator-68c6474976-6nxfz\" (UID: \"5bb19d8d-559a-4d0a-9009-f1eaf935b36b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.933942 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.941111 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b48gz\" (UniqueName: \"kubernetes.io/projected/c77eadcf-e03b-4cc7-901c-3ecd159a47bc-kube-api-access-b48gz\") pod \"package-server-manager-789f6589d5-5mp4f\" (UID: \"c77eadcf-e03b-4cc7-901c-3ecd159a47bc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.956979 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.961526 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vbgw9"] Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.966165 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpdch\" (UniqueName: \"kubernetes.io/projected/b4901cb1-78b0-4aa8-9e84-22d3c1f6669c-kube-api-access-cpdch\") pod \"downloads-7954f5f757-mfqr6\" (UID: \"b4901cb1-78b0-4aa8-9e84-22d3c1f6669c\") " pod="openshift-console/downloads-7954f5f757-mfqr6" Oct 04 08:18:34 crc kubenswrapper[4969]: I1004 08:18:34.987020 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bqtk\" (UniqueName: \"kubernetes.io/projected/846186fd-c540-464d-b85a-a67f0c8870b6-kube-api-access-2bqtk\") pod \"controller-manager-879f6c89f-lrtsv\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.008201 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx7nc\" (UniqueName: \"kubernetes.io/projected/5f8ffddc-8c22-4008-850c-b2d9a6738f88-kube-api-access-bx7nc\") pod \"cluster-samples-operator-665b6dd947-p2lsw\" (UID: \"5f8ffddc-8c22-4008-850c-b2d9a6738f88\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.021354 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9ffd\" (UniqueName: \"kubernetes.io/projected/1a654328-5c34-4927-949f-0b5818da9b49-kube-api-access-g9ffd\") pod \"console-f9d7485db-m67tq\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.033590 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.043083 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jwzk\" (UniqueName: \"kubernetes.io/projected/0be6fb94-7ce5-4556-88f7-0938db54d66e-kube-api-access-9jwzk\") pod \"machine-config-controller-84d6567774-2ksr9\" (UID: \"0be6fb94-7ce5-4556-88f7-0938db54d66e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.043384 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.049246 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mfqr6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.058837 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.062934 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f20abc5-82af-483a-a893-3855b6df7fa5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wtczq\" (UID: \"0f20abc5-82af-483a-a893-3855b6df7fa5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.064129 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.078859 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d7sh\" (UniqueName: \"kubernetes.io/projected/d075b9d0-e883-4f71-a7e6-b5397d983a8b-kube-api-access-5d7sh\") pod \"etcd-operator-b45778765-hbvs6\" (UID: \"d075b9d0-e883-4f71-a7e6-b5397d983a8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.085712 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.104830 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.106572 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppwwb\" (UniqueName: \"kubernetes.io/projected/7aa45ab6-8235-4e99-af75-c60611ae22f6-kube-api-access-ppwwb\") pod \"openshift-apiserver-operator-796bbdcf4f-qwtnd\" (UID: \"7aa45ab6-8235-4e99-af75-c60611ae22f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.115659 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.116131 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84767\" (UniqueName: \"kubernetes.io/projected/e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde-kube-api-access-84767\") pod \"dns-operator-744455d44c-xtlw2\" (UID: \"e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde\") " pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.127673 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-swbhw"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.148362 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wd65\" (UniqueName: \"kubernetes.io/projected/005c5c3d-9ede-4170-9084-b55e868faab5-kube-api-access-5wd65\") pod \"apiserver-76f77b778f-t4b7l\" (UID: \"005c5c3d-9ede-4170-9084-b55e868faab5\") " pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.160397 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mkjx\" (UniqueName: \"kubernetes.io/projected/c5ad8d46-2274-448b-a6bf-e2df545679f9-kube-api-access-5mkjx\") pod \"migrator-59844c95c7-xcfzj\" (UID: \"c5ad8d46-2274-448b-a6bf-e2df545679f9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.162526 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.177429 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.181655 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.185683 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.188842 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t2ml\" (UniqueName: \"kubernetes.io/projected/2c4fb768-95d5-4587-9e36-cc0d4ce79d5b-kube-api-access-4t2ml\") pod \"openshift-config-operator-7777fb866f-nhfgd\" (UID: \"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.211725 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.219722 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.223736 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.236284 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.267892 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-policies\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.267920 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th8lt\" (UniqueName: \"kubernetes.io/projected/d50d067c-2b69-4e82-8213-a39421d81df4-kube-api-access-th8lt\") pod \"kube-storage-version-migrator-operator-b67b599dd-45hd8\" (UID: \"d50d067c-2b69-4e82-8213-a39421d81df4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.267944 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7b66b40-6c4e-41fc-ac78-e5000ce141a5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xtfkc\" (UID: \"f7b66b40-6c4e-41fc-ac78-e5000ce141a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.267960 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.267979 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/07fc39da-f348-4892-a8ab-df8200d861af-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.267997 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4jz5\" (UniqueName: \"kubernetes.io/projected/bf40de47-36a7-470e-b59b-5eea38cbc8ed-kube-api-access-g4jz5\") pod \"service-ca-operator-777779d784-zps22\" (UID: \"bf40de47-36a7-470e-b59b-5eea38cbc8ed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.268121 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn7q4\" (UniqueName: \"kubernetes.io/projected/6379393b-2408-4b74-adf7-713fcfd448dd-kube-api-access-sn7q4\") pod \"olm-operator-6b444d44fb-k9hcj\" (UID: \"6379393b-2408-4b74-adf7-713fcfd448dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.268464 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5f151666-21ca-4ecd-ae16-6818a5780949-apiservice-cert\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.268498 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-registry-certificates\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.268516 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd62df54-4f54-45fa-bcb0-92fd3ec71631-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-p49br\" (UID: \"fd62df54-4f54-45fa-bcb0-92fd3ec71631\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269328 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269348 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn586\" (UniqueName: \"kubernetes.io/projected/d0a906da-f3d3-40cf-a714-2159253bb043-kube-api-access-nn586\") pod \"marketplace-operator-79b997595-7tnxs\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269368 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-registry-tls\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269385 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269402 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d24a7bdf-1b12-49f8-b351-896f6edf2654-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mtn99\" (UID: \"d24a7bdf-1b12-49f8-b351-896f6edf2654\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269444 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6414b732-c979-4985-a35b-f1a45c626cc2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269465 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwsdq\" (UniqueName: \"kubernetes.io/projected/d24a7bdf-1b12-49f8-b351-896f6edf2654-kube-api-access-vwsdq\") pod \"control-plane-machine-set-operator-78cbb6b69f-mtn99\" (UID: \"d24a7bdf-1b12-49f8-b351-896f6edf2654\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269481 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgmpz\" (UniqueName: \"kubernetes.io/projected/92b0b93f-7a67-4784-b776-c24b09b0774d-kube-api-access-bgmpz\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269506 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-trusted-ca\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269523 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5f151666-21ca-4ecd-ae16-6818a5780949-webhook-cert\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269543 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bht6\" (UniqueName: \"kubernetes.io/projected/29195139-26ff-4158-ba2c-282b1b921ede-kube-api-access-6bht6\") pod \"collect-profiles-29326095-4p88x\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269558 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5aaedd0-47fa-4a7a-b61d-8776444d076b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269744 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269767 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50d067c-2b69-4e82-8213-a39421d81df4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-45hd8\" (UID: \"d50d067c-2b69-4e82-8213-a39421d81df4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269792 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50d067c-2b69-4e82-8213-a39421d81df4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-45hd8\" (UID: \"d50d067c-2b69-4e82-8213-a39421d81df4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269810 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb7gj\" (UniqueName: \"kubernetes.io/projected/fcef15bf-980b-4d07-a5fb-7ab2145f69e4-kube-api-access-zb7gj\") pod \"multus-admission-controller-857f4d67dd-rc9xt\" (UID: \"fcef15bf-980b-4d07-a5fb-7ab2145f69e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269825 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f9adb985-01b8-481e-8c7a-d0e64efe0587-stats-auth\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269844 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7ad46146-89bb-43b6-badb-782f113169a9-images\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269860 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/baaa9767-dc9f-478a-9202-9ad67c7a9d56-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5mtlf\" (UID: \"baaa9767-dc9f-478a-9202-9ad67c7a9d56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269875 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjxdc\" (UniqueName: \"kubernetes.io/projected/f9adb985-01b8-481e-8c7a-d0e64efe0587-kube-api-access-pjxdc\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.269929 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9adb985-01b8-481e-8c7a-d0e64efe0587-service-ca-bundle\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.270128 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9adb985-01b8-481e-8c7a-d0e64efe0587-metrics-certs\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.270159 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwszp\" (UniqueName: \"kubernetes.io/projected/5f151666-21ca-4ecd-ae16-6818a5780949-kube-api-access-mwszp\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.270952 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5aaedd0-47fa-4a7a-b61d-8776444d076b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.271670 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0b4fed59-093e-4cc0-ad3e-00825718277b-signing-cabundle\") pod \"service-ca-9c57cc56f-8kcmp\" (UID: \"0b4fed59-093e-4cc0-ad3e-00825718277b\") " pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.271718 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.271750 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fcef15bf-980b-4d07-a5fb-7ab2145f69e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rc9xt\" (UID: \"fcef15bf-980b-4d07-a5fb-7ab2145f69e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.271769 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6414b732-c979-4985-a35b-f1a45c626cc2-metrics-tls\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.271796 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92b0b93f-7a67-4784-b776-c24b09b0774d-trusted-ca\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.271928 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnssm\" (UniqueName: \"kubernetes.io/projected/fd62df54-4f54-45fa-bcb0-92fd3ec71631-kube-api-access-vnssm\") pod \"openshift-controller-manager-operator-756b6f6bc6-p49br\" (UID: \"fd62df54-4f54-45fa-bcb0-92fd3ec71631\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.272103 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5aaedd0-47fa-4a7a-b61d-8776444d076b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.272210 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.272298 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.272325 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7ad46146-89bb-43b6-badb-782f113169a9-proxy-tls\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.272368 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.272403 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6414b732-c979-4985-a35b-f1a45c626cc2-trusted-ca\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.272629 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.273621 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfb2x\" (UniqueName: \"kubernetes.io/projected/0b4fed59-093e-4cc0-ad3e-00825718277b-kube-api-access-gfb2x\") pod \"service-ca-9c57cc56f-8kcmp\" (UID: \"0b4fed59-093e-4cc0-ad3e-00825718277b\") " pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.273689 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7tnxs\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.274009 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92b0b93f-7a67-4784-b776-c24b09b0774d-serving-cert\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.274342 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.274400 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29195139-26ff-4158-ba2c-282b1b921ede-secret-volume\") pod \"collect-profiles-29326095-4p88x\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.274455 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baaa9767-dc9f-478a-9202-9ad67c7a9d56-config\") pod \"kube-apiserver-operator-766d6c64bb-5mtlf\" (UID: \"baaa9767-dc9f-478a-9202-9ad67c7a9d56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.274859 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92b0b93f-7a67-4784-b776-c24b09b0774d-config\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.274904 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ad46146-89bb-43b6-badb-782f113169a9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.274966 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5f151666-21ca-4ecd-ae16-6818a5780949-tmpfs\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.274997 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7tnxs\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.275020 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf40de47-36a7-470e-b59b-5eea38cbc8ed-config\") pod \"service-ca-operator-777779d784-zps22\" (UID: \"bf40de47-36a7-470e-b59b-5eea38cbc8ed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.275042 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0b4fed59-093e-4cc0-ad3e-00825718277b-signing-key\") pod \"service-ca-9c57cc56f-8kcmp\" (UID: \"0b4fed59-093e-4cc0-ad3e-00825718277b\") " pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.275111 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-bound-sa-token\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.275141 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6379393b-2408-4b74-adf7-713fcfd448dd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k9hcj\" (UID: \"6379393b-2408-4b74-adf7-713fcfd448dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.275510 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmrx7\" (UniqueName: \"kubernetes.io/projected/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-kube-api-access-lmrx7\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.275716 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.275791 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrvjl\" (UniqueName: \"kubernetes.io/projected/d5aaedd0-47fa-4a7a-b61d-8776444d076b-kube-api-access-mrvjl\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.275821 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f9adb985-01b8-481e-8c7a-d0e64efe0587-default-certificate\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.275978 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:35.775967439 +0000 UTC m=+143.530236253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276139 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/baaa9767-dc9f-478a-9202-9ad67c7a9d56-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5mtlf\" (UID: \"baaa9767-dc9f-478a-9202-9ad67c7a9d56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276237 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276261 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf40de47-36a7-470e-b59b-5eea38cbc8ed-serving-cert\") pod \"service-ca-operator-777779d784-zps22\" (UID: \"bf40de47-36a7-470e-b59b-5eea38cbc8ed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276330 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-dir\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276350 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6379393b-2408-4b74-adf7-713fcfd448dd-srv-cert\") pod \"olm-operator-6b444d44fb-k9hcj\" (UID: \"6379393b-2408-4b74-adf7-713fcfd448dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276391 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t5dz\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-kube-api-access-5t5dz\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276462 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/07fc39da-f348-4892-a8ab-df8200d861af-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276502 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcq76\" (UniqueName: \"kubernetes.io/projected/7ad46146-89bb-43b6-badb-782f113169a9-kube-api-access-kcq76\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276523 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd62df54-4f54-45fa-bcb0-92fd3ec71631-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-p49br\" (UID: \"fd62df54-4f54-45fa-bcb0-92fd3ec71631\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276864 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7b66b40-6c4e-41fc-ac78-e5000ce141a5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xtfkc\" (UID: \"f7b66b40-6c4e-41fc-ac78-e5000ce141a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.276892 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29195139-26ff-4158-ba2c-282b1b921ede-config-volume\") pod \"collect-profiles-29326095-4p88x\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.277102 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wzb5\" (UniqueName: \"kubernetes.io/projected/6414b732-c979-4985-a35b-f1a45c626cc2-kube-api-access-4wzb5\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.277484 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7b66b40-6c4e-41fc-ac78-e5000ce141a5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xtfkc\" (UID: \"f7b66b40-6c4e-41fc-ac78-e5000ce141a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.328617 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.374853 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378447 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.378599 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:35.87857696 +0000 UTC m=+143.632845774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378674 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f253261-82ae-49f4-9bf7-0598509f1cb2-cert\") pod \"ingress-canary-2bhvf\" (UID: \"1f253261-82ae-49f4-9bf7-0598509f1cb2\") " pod="openshift-ingress-canary/ingress-canary-2bhvf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378758 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378780 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7932b6aa-050e-4f49-bb0f-82ec7575d452-config-volume\") pod \"dns-default-tmqwr\" (UID: \"7932b6aa-050e-4f49-bb0f-82ec7575d452\") " pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378816 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfb2x\" (UniqueName: \"kubernetes.io/projected/0b4fed59-093e-4cc0-ad3e-00825718277b-kube-api-access-gfb2x\") pod \"service-ca-9c57cc56f-8kcmp\" (UID: \"0b4fed59-093e-4cc0-ad3e-00825718277b\") " pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378836 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-socket-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378853 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7tnxs\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378874 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-mountpoint-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378944 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92b0b93f-7a67-4784-b776-c24b09b0774d-serving-cert\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.378999 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29195139-26ff-4158-ba2c-282b1b921ede-secret-volume\") pod \"collect-profiles-29326095-4p88x\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379017 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baaa9767-dc9f-478a-9202-9ad67c7a9d56-config\") pod \"kube-apiserver-operator-766d6c64bb-5mtlf\" (UID: \"baaa9767-dc9f-478a-9202-9ad67c7a9d56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379068 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379086 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92b0b93f-7a67-4784-b776-c24b09b0774d-config\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379104 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ad46146-89bb-43b6-badb-782f113169a9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379175 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5f151666-21ca-4ecd-ae16-6818a5780949-tmpfs\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379195 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7tnxs\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379229 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf40de47-36a7-470e-b59b-5eea38cbc8ed-config\") pod \"service-ca-operator-777779d784-zps22\" (UID: \"bf40de47-36a7-470e-b59b-5eea38cbc8ed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379246 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-bound-sa-token\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379263 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6379393b-2408-4b74-adf7-713fcfd448dd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k9hcj\" (UID: \"6379393b-2408-4b74-adf7-713fcfd448dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379277 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0b4fed59-093e-4cc0-ad3e-00825718277b-signing-key\") pod \"service-ca-9c57cc56f-8kcmp\" (UID: \"0b4fed59-093e-4cc0-ad3e-00825718277b\") " pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379324 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmrx7\" (UniqueName: \"kubernetes.io/projected/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-kube-api-access-lmrx7\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379399 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379473 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrvjl\" (UniqueName: \"kubernetes.io/projected/d5aaedd0-47fa-4a7a-b61d-8776444d076b-kube-api-access-mrvjl\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379524 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/baaa9767-dc9f-478a-9202-9ad67c7a9d56-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5mtlf\" (UID: \"baaa9767-dc9f-478a-9202-9ad67c7a9d56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379541 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f9adb985-01b8-481e-8c7a-d0e64efe0587-default-certificate\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379560 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxkc7\" (UniqueName: \"kubernetes.io/projected/7932b6aa-050e-4f49-bb0f-82ec7575d452-kube-api-access-wxkc7\") pod \"dns-default-tmqwr\" (UID: \"7932b6aa-050e-4f49-bb0f-82ec7575d452\") " pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379578 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-plugins-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379595 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379610 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf40de47-36a7-470e-b59b-5eea38cbc8ed-serving-cert\") pod \"service-ca-operator-777779d784-zps22\" (UID: \"bf40de47-36a7-470e-b59b-5eea38cbc8ed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379625 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-dir\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.379643 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6379393b-2408-4b74-adf7-713fcfd448dd-srv-cert\") pod \"olm-operator-6b444d44fb-k9hcj\" (UID: \"6379393b-2408-4b74-adf7-713fcfd448dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.380376 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf40de47-36a7-470e-b59b-5eea38cbc8ed-config\") pod \"service-ca-operator-777779d784-zps22\" (UID: \"bf40de47-36a7-470e-b59b-5eea38cbc8ed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.381113 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7tnxs\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.381156 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t5dz\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-kube-api-access-5t5dz\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.381175 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/07fc39da-f348-4892-a8ab-df8200d861af-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.381195 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcq76\" (UniqueName: \"kubernetes.io/projected/7ad46146-89bb-43b6-badb-782f113169a9-kube-api-access-kcq76\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.381219 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c-certs\") pod \"machine-config-server-4qbp6\" (UID: \"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c\") " pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.381239 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5bpn\" (UniqueName: \"kubernetes.io/projected/1f253261-82ae-49f4-9bf7-0598509f1cb2-kube-api-access-q5bpn\") pod \"ingress-canary-2bhvf\" (UID: \"1f253261-82ae-49f4-9bf7-0598509f1cb2\") " pod="openshift-ingress-canary/ingress-canary-2bhvf" Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.381246 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:35.881236577 +0000 UTC m=+143.635505391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.381997 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92b0b93f-7a67-4784-b776-c24b09b0774d-config\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.382266 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.382326 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ad46146-89bb-43b6-badb-782f113169a9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.382909 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baaa9767-dc9f-478a-9202-9ad67c7a9d56-config\") pod \"kube-apiserver-operator-766d6c64bb-5mtlf\" (UID: \"baaa9767-dc9f-478a-9202-9ad67c7a9d56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.386026 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5f151666-21ca-4ecd-ae16-6818a5780949-tmpfs\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.386149 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0b4fed59-093e-4cc0-ad3e-00825718277b-signing-key\") pod \"service-ca-9c57cc56f-8kcmp\" (UID: \"0b4fed59-093e-4cc0-ad3e-00825718277b\") " pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.386204 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-dir\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.386256 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd62df54-4f54-45fa-bcb0-92fd3ec71631-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-p49br\" (UID: \"fd62df54-4f54-45fa-bcb0-92fd3ec71631\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.386402 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-csi-data-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.386654 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7b66b40-6c4e-41fc-ac78-e5000ce141a5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xtfkc\" (UID: \"f7b66b40-6c4e-41fc-ac78-e5000ce141a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.387018 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd62df54-4f54-45fa-bcb0-92fd3ec71631-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-p49br\" (UID: \"fd62df54-4f54-45fa-bcb0-92fd3ec71631\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.387150 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29195139-26ff-4158-ba2c-282b1b921ede-config-volume\") pod \"collect-profiles-29326095-4p88x\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.387211 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-registration-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.387888 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29195139-26ff-4158-ba2c-282b1b921ede-config-volume\") pod \"collect-profiles-29326095-4p88x\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.387953 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wzb5\" (UniqueName: \"kubernetes.io/projected/6414b732-c979-4985-a35b-f1a45c626cc2-kube-api-access-4wzb5\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388131 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7b66b40-6c4e-41fc-ac78-e5000ce141a5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xtfkc\" (UID: \"f7b66b40-6c4e-41fc-ac78-e5000ce141a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388241 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r5l5\" (UniqueName: \"kubernetes.io/projected/020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c-kube-api-access-2r5l5\") pod \"machine-config-server-4qbp6\" (UID: \"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c\") " pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388362 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-policies\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388403 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th8lt\" (UniqueName: \"kubernetes.io/projected/d50d067c-2b69-4e82-8213-a39421d81df4-kube-api-access-th8lt\") pod \"kube-storage-version-migrator-operator-b67b599dd-45hd8\" (UID: \"d50d067c-2b69-4e82-8213-a39421d81df4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388447 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7b66b40-6c4e-41fc-ac78-e5000ce141a5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xtfkc\" (UID: \"f7b66b40-6c4e-41fc-ac78-e5000ce141a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388500 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/07fc39da-f348-4892-a8ab-df8200d861af-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388520 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388540 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4jz5\" (UniqueName: \"kubernetes.io/projected/bf40de47-36a7-470e-b59b-5eea38cbc8ed-kube-api-access-g4jz5\") pod \"service-ca-operator-777779d784-zps22\" (UID: \"bf40de47-36a7-470e-b59b-5eea38cbc8ed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388556 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5f151666-21ca-4ecd-ae16-6818a5780949-apiservice-cert\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388574 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn7q4\" (UniqueName: \"kubernetes.io/projected/6379393b-2408-4b74-adf7-713fcfd448dd-kube-api-access-sn7q4\") pod \"olm-operator-6b444d44fb-k9hcj\" (UID: \"6379393b-2408-4b74-adf7-713fcfd448dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388592 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd62df54-4f54-45fa-bcb0-92fd3ec71631-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-p49br\" (UID: \"fd62df54-4f54-45fa-bcb0-92fd3ec71631\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388608 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-registry-certificates\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388634 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn586\" (UniqueName: \"kubernetes.io/projected/d0a906da-f3d3-40cf-a714-2159253bb043-kube-api-access-nn586\") pod \"marketplace-operator-79b997595-7tnxs\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388651 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388670 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d24a7bdf-1b12-49f8-b351-896f6edf2654-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mtn99\" (UID: \"d24a7bdf-1b12-49f8-b351-896f6edf2654\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388686 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6414b732-c979-4985-a35b-f1a45c626cc2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388702 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-registry-tls\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388718 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388743 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwsdq\" (UniqueName: \"kubernetes.io/projected/d24a7bdf-1b12-49f8-b351-896f6edf2654-kube-api-access-vwsdq\") pod \"control-plane-machine-set-operator-78cbb6b69f-mtn99\" (UID: \"d24a7bdf-1b12-49f8-b351-896f6edf2654\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388758 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgmpz\" (UniqueName: \"kubernetes.io/projected/92b0b93f-7a67-4784-b776-c24b09b0774d-kube-api-access-bgmpz\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388783 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-trusted-ca\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388800 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5f151666-21ca-4ecd-ae16-6818a5780949-webhook-cert\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388816 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bht6\" (UniqueName: \"kubernetes.io/projected/29195139-26ff-4158-ba2c-282b1b921ede-kube-api-access-6bht6\") pod \"collect-profiles-29326095-4p88x\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388833 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5aaedd0-47fa-4a7a-b61d-8776444d076b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388851 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388869 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50d067c-2b69-4e82-8213-a39421d81df4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-45hd8\" (UID: \"d50d067c-2b69-4e82-8213-a39421d81df4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388896 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7932b6aa-050e-4f49-bb0f-82ec7575d452-metrics-tls\") pod \"dns-default-tmqwr\" (UID: \"7932b6aa-050e-4f49-bb0f-82ec7575d452\") " pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388915 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c-node-bootstrap-token\") pod \"machine-config-server-4qbp6\" (UID: \"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c\") " pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388933 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50d067c-2b69-4e82-8213-a39421d81df4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-45hd8\" (UID: \"d50d067c-2b69-4e82-8213-a39421d81df4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388950 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb7gj\" (UniqueName: \"kubernetes.io/projected/fcef15bf-980b-4d07-a5fb-7ab2145f69e4-kube-api-access-zb7gj\") pod \"multus-admission-controller-857f4d67dd-rc9xt\" (UID: \"fcef15bf-980b-4d07-a5fb-7ab2145f69e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388965 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f9adb985-01b8-481e-8c7a-d0e64efe0587-stats-auth\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388979 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjxdc\" (UniqueName: \"kubernetes.io/projected/f9adb985-01b8-481e-8c7a-d0e64efe0587-kube-api-access-pjxdc\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.388996 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7ad46146-89bb-43b6-badb-782f113169a9-images\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389012 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/baaa9767-dc9f-478a-9202-9ad67c7a9d56-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5mtlf\" (UID: \"baaa9767-dc9f-478a-9202-9ad67c7a9d56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389043 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9adb985-01b8-481e-8c7a-d0e64efe0587-service-ca-bundle\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389061 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9adb985-01b8-481e-8c7a-d0e64efe0587-metrics-certs\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389088 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwszp\" (UniqueName: \"kubernetes.io/projected/5f151666-21ca-4ecd-ae16-6818a5780949-kube-api-access-mwszp\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389113 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5aaedd0-47fa-4a7a-b61d-8776444d076b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389140 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389155 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fcef15bf-980b-4d07-a5fb-7ab2145f69e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rc9xt\" (UID: \"fcef15bf-980b-4d07-a5fb-7ab2145f69e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389170 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0b4fed59-093e-4cc0-ad3e-00825718277b-signing-cabundle\") pod \"service-ca-9c57cc56f-8kcmp\" (UID: \"0b4fed59-093e-4cc0-ad3e-00825718277b\") " pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389186 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg777\" (UniqueName: \"kubernetes.io/projected/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-kube-api-access-gg777\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389212 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6414b732-c979-4985-a35b-f1a45c626cc2-metrics-tls\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389226 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92b0b93f-7a67-4784-b776-c24b09b0774d-trusted-ca\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389253 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnssm\" (UniqueName: \"kubernetes.io/projected/fd62df54-4f54-45fa-bcb0-92fd3ec71631-kube-api-access-vnssm\") pod \"openshift-controller-manager-operator-756b6f6bc6-p49br\" (UID: \"fd62df54-4f54-45fa-bcb0-92fd3ec71631\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389277 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5aaedd0-47fa-4a7a-b61d-8776444d076b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389296 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389322 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389348 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7ad46146-89bb-43b6-badb-782f113169a9-proxy-tls\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389365 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.389381 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6414b732-c979-4985-a35b-f1a45c626cc2-trusted-ca\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.391442 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7b66b40-6c4e-41fc-ac78-e5000ce141a5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xtfkc\" (UID: \"f7b66b40-6c4e-41fc-ac78-e5000ce141a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.392586 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.394402 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5aaedd0-47fa-4a7a-b61d-8776444d076b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.394392 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92b0b93f-7a67-4784-b776-c24b09b0774d-serving-cert\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.394857 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/07fc39da-f348-4892-a8ab-df8200d861af-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.394878 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0b4fed59-093e-4cc0-ad3e-00825718277b-signing-cabundle\") pod \"service-ca-9c57cc56f-8kcmp\" (UID: \"0b4fed59-093e-4cc0-ad3e-00825718277b\") " pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.395908 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-policies\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.396225 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7ad46146-89bb-43b6-badb-782f113169a9-images\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.396531 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7b66b40-6c4e-41fc-ac78-e5000ce141a5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xtfkc\" (UID: \"f7b66b40-6c4e-41fc-ac78-e5000ce141a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.396683 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-registry-tls\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.396728 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6379393b-2408-4b74-adf7-713fcfd448dd-srv-cert\") pod \"olm-operator-6b444d44fb-k9hcj\" (UID: \"6379393b-2408-4b74-adf7-713fcfd448dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.396832 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92b0b93f-7a67-4784-b776-c24b09b0774d-trusted-ca\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.397069 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7tnxs\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.397160 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9adb985-01b8-481e-8c7a-d0e64efe0587-service-ca-bundle\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.397503 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fcef15bf-980b-4d07-a5fb-7ab2145f69e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rc9xt\" (UID: \"fcef15bf-980b-4d07-a5fb-7ab2145f69e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.397752 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-trusted-ca\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.397985 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-registry-certificates\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.398130 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.398230 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50d067c-2b69-4e82-8213-a39421d81df4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-45hd8\" (UID: \"d50d067c-2b69-4e82-8213-a39421d81df4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.399091 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.405036 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.405492 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29195139-26ff-4158-ba2c-282b1b921ede-secret-volume\") pod \"collect-profiles-29326095-4p88x\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.405686 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf40de47-36a7-470e-b59b-5eea38cbc8ed-serving-cert\") pod \"service-ca-operator-777779d784-zps22\" (UID: \"bf40de47-36a7-470e-b59b-5eea38cbc8ed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.405779 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7ad46146-89bb-43b6-badb-782f113169a9-proxy-tls\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.405865 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.406106 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6379393b-2408-4b74-adf7-713fcfd448dd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k9hcj\" (UID: \"6379393b-2408-4b74-adf7-713fcfd448dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.406706 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50d067c-2b69-4e82-8213-a39421d81df4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-45hd8\" (UID: \"d50d067c-2b69-4e82-8213-a39421d81df4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.406769 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6414b732-c979-4985-a35b-f1a45c626cc2-trusted-ca\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.407572 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/07fc39da-f348-4892-a8ab-df8200d861af-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.408105 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f9adb985-01b8-481e-8c7a-d0e64efe0587-default-certificate\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.414194 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5f151666-21ca-4ecd-ae16-6818a5780949-apiservice-cert\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.414583 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.415199 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5aaedd0-47fa-4a7a-b61d-8776444d076b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.416770 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd62df54-4f54-45fa-bcb0-92fd3ec71631-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-p49br\" (UID: \"fd62df54-4f54-45fa-bcb0-92fd3ec71631\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.417194 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.418848 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5f151666-21ca-4ecd-ae16-6818a5780949-webhook-cert\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.419218 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/baaa9767-dc9f-478a-9202-9ad67c7a9d56-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5mtlf\" (UID: \"baaa9767-dc9f-478a-9202-9ad67c7a9d56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.419375 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f9adb985-01b8-481e-8c7a-d0e64efe0587-stats-auth\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.420376 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6414b732-c979-4985-a35b-f1a45c626cc2-metrics-tls\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.420451 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.420750 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9adb985-01b8-481e-8c7a-d0e64efe0587-metrics-certs\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.420915 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d24a7bdf-1b12-49f8-b351-896f6edf2654-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mtn99\" (UID: \"d24a7bdf-1b12-49f8-b351-896f6edf2654\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.421667 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.421792 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.422396 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.424704 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmrx7\" (UniqueName: \"kubernetes.io/projected/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-kube-api-access-lmrx7\") pod \"oauth-openshift-558db77b4-8xngh\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.439029 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-bound-sa-token\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.458556 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcq76\" (UniqueName: \"kubernetes.io/projected/7ad46146-89bb-43b6-badb-782f113169a9-kube-api-access-kcq76\") pod \"machine-config-operator-74547568cd-h8gcd\" (UID: \"7ad46146-89bb-43b6-badb-782f113169a9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.497201 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t5dz\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-kube-api-access-5t5dz\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499039 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499161 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f253261-82ae-49f4-9bf7-0598509f1cb2-cert\") pod \"ingress-canary-2bhvf\" (UID: \"1f253261-82ae-49f4-9bf7-0598509f1cb2\") " pod="openshift-ingress-canary/ingress-canary-2bhvf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499178 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7932b6aa-050e-4f49-bb0f-82ec7575d452-config-volume\") pod \"dns-default-tmqwr\" (UID: \"7932b6aa-050e-4f49-bb0f-82ec7575d452\") " pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499202 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-socket-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499216 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-mountpoint-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499269 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxkc7\" (UniqueName: \"kubernetes.io/projected/7932b6aa-050e-4f49-bb0f-82ec7575d452-kube-api-access-wxkc7\") pod \"dns-default-tmqwr\" (UID: \"7932b6aa-050e-4f49-bb0f-82ec7575d452\") " pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499285 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-plugins-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499303 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c-certs\") pod \"machine-config-server-4qbp6\" (UID: \"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c\") " pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499317 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5bpn\" (UniqueName: \"kubernetes.io/projected/1f253261-82ae-49f4-9bf7-0598509f1cb2-kube-api-access-q5bpn\") pod \"ingress-canary-2bhvf\" (UID: \"1f253261-82ae-49f4-9bf7-0598509f1cb2\") " pod="openshift-ingress-canary/ingress-canary-2bhvf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499335 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-csi-data-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499352 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-registration-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499381 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r5l5\" (UniqueName: \"kubernetes.io/projected/020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c-kube-api-access-2r5l5\") pod \"machine-config-server-4qbp6\" (UID: \"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c\") " pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499499 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c-node-bootstrap-token\") pod \"machine-config-server-4qbp6\" (UID: \"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c\") " pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499515 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7932b6aa-050e-4f49-bb0f-82ec7575d452-metrics-tls\") pod \"dns-default-tmqwr\" (UID: \"7932b6aa-050e-4f49-bb0f-82ec7575d452\") " pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.499568 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg777\" (UniqueName: \"kubernetes.io/projected/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-kube-api-access-gg777\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.501265 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-plugins-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.501400 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.001382089 +0000 UTC m=+143.755650903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.501658 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-registration-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.503357 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-mountpoint-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.503414 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-csi-data-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.503838 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrvjl\" (UniqueName: \"kubernetes.io/projected/d5aaedd0-47fa-4a7a-b61d-8776444d076b-kube-api-access-mrvjl\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.503947 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7932b6aa-050e-4f49-bb0f-82ec7575d452-config-volume\") pod \"dns-default-tmqwr\" (UID: \"7932b6aa-050e-4f49-bb0f-82ec7575d452\") " pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.504193 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-socket-dir\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.505949 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c-certs\") pod \"machine-config-server-4qbp6\" (UID: \"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c\") " pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.508035 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7932b6aa-050e-4f49-bb0f-82ec7575d452-metrics-tls\") pod \"dns-default-tmqwr\" (UID: \"7932b6aa-050e-4f49-bb0f-82ec7575d452\") " pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.510480 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c-node-bootstrap-token\") pod \"machine-config-server-4qbp6\" (UID: \"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c\") " pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.511470 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f253261-82ae-49f4-9bf7-0598509f1cb2-cert\") pod \"ingress-canary-2bhvf\" (UID: \"1f253261-82ae-49f4-9bf7-0598509f1cb2\") " pod="openshift-ingress-canary/ingress-canary-2bhvf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.520405 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mfqr6"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.529008 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfb2x\" (UniqueName: \"kubernetes.io/projected/0b4fed59-093e-4cc0-ad3e-00825718277b-kube-api-access-gfb2x\") pod \"service-ca-9c57cc56f-8kcmp\" (UID: \"0b4fed59-093e-4cc0-ad3e-00825718277b\") " pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.530702 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.545504 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/baaa9767-dc9f-478a-9202-9ad67c7a9d56-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5mtlf\" (UID: \"baaa9767-dc9f-478a-9202-9ad67c7a9d56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.556783 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wzb5\" (UniqueName: \"kubernetes.io/projected/6414b732-c979-4985-a35b-f1a45c626cc2-kube-api-access-4wzb5\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.579157 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.602163 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.602729 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.102715958 +0000 UTC m=+143.856984772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.628585 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.629669 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7b66b40-6c4e-41fc-ac78-e5000ce141a5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xtfkc\" (UID: \"f7b66b40-6c4e-41fc-ac78-e5000ce141a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.636656 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn7q4\" (UniqueName: \"kubernetes.io/projected/6379393b-2408-4b74-adf7-713fcfd448dd-kube-api-access-sn7q4\") pod \"olm-operator-6b444d44fb-k9hcj\" (UID: \"6379393b-2408-4b74-adf7-713fcfd448dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.645788 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.648801 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th8lt\" (UniqueName: \"kubernetes.io/projected/d50d067c-2b69-4e82-8213-a39421d81df4-kube-api-access-th8lt\") pod \"kube-storage-version-migrator-operator-b67b599dd-45hd8\" (UID: \"d50d067c-2b69-4e82-8213-a39421d81df4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.654458 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.671504 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6414b732-c979-4985-a35b-f1a45c626cc2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ddzpn\" (UID: \"6414b732-c979-4985-a35b-f1a45c626cc2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.688283 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4jz5\" (UniqueName: \"kubernetes.io/projected/bf40de47-36a7-470e-b59b-5eea38cbc8ed-kube-api-access-g4jz5\") pod \"service-ca-operator-777779d784-zps22\" (UID: \"bf40de47-36a7-470e-b59b-5eea38cbc8ed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.699391 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.702495 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn586\" (UniqueName: \"kubernetes.io/projected/d0a906da-f3d3-40cf-a714-2159253bb043-kube-api-access-nn586\") pod \"marketplace-operator-79b997595-7tnxs\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.702925 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.706557 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.206530999 +0000 UTC m=+143.960799813 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.706776 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.707204 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.207192306 +0000 UTC m=+143.961461120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.721248 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb7gj\" (UniqueName: \"kubernetes.io/projected/fcef15bf-980b-4d07-a5fb-7ab2145f69e4-kube-api-access-zb7gj\") pod \"multus-admission-controller-857f4d67dd-rc9xt\" (UID: \"fcef15bf-980b-4d07-a5fb-7ab2145f69e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.737859 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5aaedd0-47fa-4a7a-b61d-8776444d076b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jt9z4\" (UID: \"d5aaedd0-47fa-4a7a-b61d-8776444d076b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.760459 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m67tq"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.761216 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t4b7l"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.769731 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.774873 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjxdc\" (UniqueName: \"kubernetes.io/projected/f9adb985-01b8-481e-8c7a-d0e64efe0587-kube-api-access-pjxdc\") pod \"router-default-5444994796-kxtxq\" (UID: \"f9adb985-01b8-481e-8c7a-d0e64efe0587\") " pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.775731 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.780225 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnssm\" (UniqueName: \"kubernetes.io/projected/fd62df54-4f54-45fa-bcb0-92fd3ec71631-kube-api-access-vnssm\") pod \"openshift-controller-manager-operator-756b6f6bc6-p49br\" (UID: \"fd62df54-4f54-45fa-bcb0-92fd3ec71631\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: W1004 08:18:35.791705 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod005c5c3d_9ede_4170_9084_b55e868faab5.slice/crio-37f00aa3b74838b2b779fe8f7a4f25918b729c88582e70ddb3b4dab512da391f WatchSource:0}: Error finding container 37f00aa3b74838b2b779fe8f7a4f25918b729c88582e70ddb3b4dab512da391f: Status 404 returned error can't find the container with id 37f00aa3b74838b2b779fe8f7a4f25918b729c88582e70ddb3b4dab512da391f Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.793975 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.797725 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.803252 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.805361 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hbvs6"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.811207 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.811595 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.311572962 +0000 UTC m=+144.065841776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.811801 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.812468 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwsdq\" (UniqueName: \"kubernetes.io/projected/d24a7bdf-1b12-49f8-b351-896f6edf2654-kube-api-access-vwsdq\") pod \"control-plane-machine-set-operator-78cbb6b69f-mtn99\" (UID: \"d24a7bdf-1b12-49f8-b351-896f6edf2654\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.814630 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.314604809 +0000 UTC m=+144.068873623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.822559 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.826354 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.826916 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgmpz\" (UniqueName: \"kubernetes.io/projected/92b0b93f-7a67-4784-b776-c24b09b0774d-kube-api-access-bgmpz\") pod \"console-operator-58897d9998-mglbz\" (UID: \"92b0b93f-7a67-4784-b776-c24b09b0774d\") " pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.837340 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lrtsv"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.839617 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.841279 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" event={"ID":"c77eadcf-e03b-4cc7-901c-3ecd159a47bc","Type":"ContainerStarted","Data":"3d7bb24439b427aece327fb909c2f28eb3ef19ab6e62fe3e96b3ae1efcb517b7"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.841304 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" event={"ID":"c77eadcf-e03b-4cc7-901c-3ecd159a47bc","Type":"ContainerStarted","Data":"27eb3974f5de5465405b9cf65c5dfecab8f080a7d82dc2547fdf9fda70a5cbd9"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.845660 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bht6\" (UniqueName: \"kubernetes.io/projected/29195139-26ff-4158-ba2c-282b1b921ede-kube-api-access-6bht6\") pod \"collect-profiles-29326095-4p88x\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.852312 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" event={"ID":"d632f337-1d3d-42c0-b1b4-3232ed684da4","Type":"ContainerStarted","Data":"efa9b3d0d0ca106f9e01c1afba82bff2b49a2f2a9ce780096a5b4d83126fabbc"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.852828 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" event={"ID":"d632f337-1d3d-42c0-b1b4-3232ed684da4","Type":"ContainerStarted","Data":"c4250552667b8108868a314d1cecea94868639beaa6c52023552f7a1a3400899"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.860507 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mfqr6" event={"ID":"b4901cb1-78b0-4aa8-9e84-22d3c1f6669c","Type":"ContainerStarted","Data":"ce7b4910820c72a20aca85184316807624f4716349ef1bcd1752a8932277975c"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.860547 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mfqr6" event={"ID":"b4901cb1-78b0-4aa8-9e84-22d3c1f6669c","Type":"ContainerStarted","Data":"93dcec762fe0aff6ca2b4c7537effe47da949076792dc346230af5df3cbac4b3"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.861163 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mfqr6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.864280 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwszp\" (UniqueName: \"kubernetes.io/projected/5f151666-21ca-4ecd-ae16-6818a5780949-kube-api-access-mwszp\") pod \"packageserver-d55dfcdfc-56p7g\" (UID: \"5f151666-21ca-4ecd-ae16-6818a5780949\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.868742 4969 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfqr6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.868784 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mfqr6" podUID="b4901cb1-78b0-4aa8-9e84-22d3c1f6669c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.875913 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" event={"ID":"5bb19d8d-559a-4d0a-9009-f1eaf935b36b","Type":"ContainerStarted","Data":"6a8de80d242760a0da6c4480e54fad7554911c1b5cf5f5b7eae5e1a5c39d4614"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.875950 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" event={"ID":"5bb19d8d-559a-4d0a-9009-f1eaf935b36b","Type":"ContainerStarted","Data":"9e52842bd5aaad243f022d9bc021014ddd831bb7875f4b6167aaaed1f25449bb"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.878980 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r5l5\" (UniqueName: \"kubernetes.io/projected/020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c-kube-api-access-2r5l5\") pod \"machine-config-server-4qbp6\" (UID: \"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c\") " pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.879031 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.879093 4969 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6nxfz container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.879127 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" podUID="5bb19d8d-559a-4d0a-9009-f1eaf935b36b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.885113 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.885215 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" event={"ID":"0703cb4b-c7c0-4632-9974-4294afec2b08","Type":"ContainerStarted","Data":"a05c05377f72f12523284e41165210109de5a2a544a53a5819546cbdd57e9eb2"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.885244 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" event={"ID":"0703cb4b-c7c0-4632-9974-4294afec2b08","Type":"ContainerStarted","Data":"2387d42aeb47b557d43fff36ad3d8aaf9fb2ea12e36c8ac40460e6b931b95167"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.885417 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xtlw2"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.889270 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m67tq" event={"ID":"1a654328-5c34-4927-949f-0b5818da9b49","Type":"ContainerStarted","Data":"b36c16b280146cdb6ba3655fb90d48769c9017b4363badd79ceae11424cd653b"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.894835 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.898935 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" event={"ID":"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8","Type":"ContainerStarted","Data":"27b64007ca90df86955eca764bd22cce34e0078d0773fef8bb9aa4a183961a50"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.898966 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" event={"ID":"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8","Type":"ContainerStarted","Data":"99d748e7f4d22ba4b8594044000e66de70f0cb3f2615fe1388d354d0ba66c096"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.898978 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" event={"ID":"5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8","Type":"ContainerStarted","Data":"c4986b5d504d58fb465db91282e1928d487d20c95ec0db2fb82c717ebd50b759"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.900217 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" event={"ID":"d7581c21-cae9-4c2b-8a46-3c08bcd54180","Type":"ContainerStarted","Data":"1f4560b83a48b765e3585ff0eeb73b63063715c87d45f30a53142cad0917f303"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.903368 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg777\" (UniqueName: \"kubernetes.io/projected/d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e-kube-api-access-gg777\") pod \"csi-hostpathplugin-x9p72\" (UID: \"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e\") " pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.912538 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.912991 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:35 crc kubenswrapper[4969]: E1004 08:18:35.913311 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.413295761 +0000 UTC m=+144.167564575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.915413 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9"] Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.916147 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" event={"ID":"005c5c3d-9ede-4170-9084-b55e868faab5","Type":"ContainerStarted","Data":"37f00aa3b74838b2b779fe8f7a4f25918b729c88582e70ddb3b4dab512da391f"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.919527 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:35 crc kubenswrapper[4969]: W1004 08:18:35.921522 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0188a25_0cb4_4e3e_8b9b_1a1a1dc47bde.slice/crio-5c9f5a36e315f0dd05ead1f5b29cc98011dadaf8a6d7ee882547cc0ef1fee344 WatchSource:0}: Error finding container 5c9f5a36e315f0dd05ead1f5b29cc98011dadaf8a6d7ee882547cc0ef1fee344: Status 404 returned error can't find the container with id 5c9f5a36e315f0dd05ead1f5b29cc98011dadaf8a6d7ee882547cc0ef1fee344 Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.924400 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" event={"ID":"12536a40-76f7-448a-a0f5-0d79f111da6b","Type":"ContainerStarted","Data":"1934722c95b7c6ef0d520e5873e2028a52a433954b2b32cc8fbd4c75d056b534"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.924627 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" event={"ID":"12536a40-76f7-448a-a0f5-0d79f111da6b","Type":"ContainerStarted","Data":"12eb8fd7d9061ae238e6d36e29bbcc23d0b01e29f44d8ae22deb2bc76798aeb7"} Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.924496 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5bpn\" (UniqueName: \"kubernetes.io/projected/1f253261-82ae-49f4-9bf7-0598509f1cb2-kube-api-access-q5bpn\") pod \"ingress-canary-2bhvf\" (UID: \"1f253261-82ae-49f4-9bf7-0598509f1cb2\") " pod="openshift-ingress-canary/ingress-canary-2bhvf" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.933562 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.942333 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxkc7\" (UniqueName: \"kubernetes.io/projected/7932b6aa-050e-4f49-bb0f-82ec7575d452-kube-api-access-wxkc7\") pod \"dns-default-tmqwr\" (UID: \"7932b6aa-050e-4f49-bb0f-82ec7575d452\") " pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.963494 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.974858 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.982655 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:35 crc kubenswrapper[4969]: I1004 08:18:35.995471 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2bhvf" Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.000674 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.008310 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4qbp6" Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.015125 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.020616 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.520601271 +0000 UTC m=+144.274870085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.028969 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-x9p72" Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.037699 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8kcmp"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.079982 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.104974 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.116495 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.117014 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.616997607 +0000 UTC m=+144.371266421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.182763 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8xngh"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.183609 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.218088 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.219581 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.719570007 +0000 UTC m=+144.473838821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.276049 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.303813 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.305042 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.322050 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.322397 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.822383634 +0000 UTC m=+144.576652448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: W1004 08:18:36.333616 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ad46146_89bb_43b6_badb_782f113169a9.slice/crio-449e150fc437ee504c64edd905924ace789f07c0af92c5f54828023052c2c87f WatchSource:0}: Error finding container 449e150fc437ee504c64edd905924ace789f07c0af92c5f54828023052c2c87f: Status 404 returned error can't find the container with id 449e150fc437ee504c64edd905924ace789f07c0af92c5f54828023052c2c87f Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.353233 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.422529 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zps22"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.423877 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.424236 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:36.924222216 +0000 UTC m=+144.678491030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.499985 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.526738 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.527022 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.027007282 +0000 UTC m=+144.781276096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.538329 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.580513 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rc9xt"] Oct 04 08:18:36 crc kubenswrapper[4969]: W1004 08:18:36.613971 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7b66b40_6c4e_41fc_ac78_e5000ce141a5.slice/crio-ecd75ecc8917cec45aa666595a8baa6d9f1daaabdbab0af58cf6f9d0128faa69 WatchSource:0}: Error finding container ecd75ecc8917cec45aa666595a8baa6d9f1daaabdbab0af58cf6f9d0128faa69: Status 404 returned error can't find the container with id ecd75ecc8917cec45aa666595a8baa6d9f1daaabdbab0af58cf6f9d0128faa69 Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.627662 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.628142 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.127962233 +0000 UTC m=+144.882231047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.679120 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.728304 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.729400 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.229372014 +0000 UTC m=+144.983640828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.831293 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.831680 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.331669247 +0000 UTC m=+145.085938061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.874722 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.932320 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:36 crc kubenswrapper[4969]: E1004 08:18:36.932993 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.432979366 +0000 UTC m=+145.187248180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.936047 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" event={"ID":"d5aaedd0-47fa-4a7a-b61d-8776444d076b","Type":"ContainerStarted","Data":"f15c9a8e9c9043cb740f727df78a81d80207900aa730bcbaca9d5f57609a2ffd"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.945350 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" event={"ID":"e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde","Type":"ContainerStarted","Data":"786271dadd77c70c9b4bfdac17be9acdbe7d3438012866d883f5476dd3720859"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.945508 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" event={"ID":"e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde","Type":"ContainerStarted","Data":"5c9f5a36e315f0dd05ead1f5b29cc98011dadaf8a6d7ee882547cc0ef1fee344"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.952404 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4qbp6" event={"ID":"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c","Type":"ContainerStarted","Data":"e44706562d8b06cb663ff68f76d757a08c44664e12607d6aec80b213139e91bf"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.973522 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7tnxs"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.983492 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" event={"ID":"0f20abc5-82af-483a-a893-3855b6df7fa5","Type":"ContainerStarted","Data":"a1b12d70c4521d44046a3d4ea86cb26efa4d843a104d9099b1b4c64dc6685c25"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.983534 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" event={"ID":"0f20abc5-82af-483a-a893-3855b6df7fa5","Type":"ContainerStarted","Data":"f372d692b91b32b601a64de177137f8cd4d3f1f3ce07b2bbdcdd95c89a060485"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.986083 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" event={"ID":"5f8ffddc-8c22-4008-850c-b2d9a6738f88","Type":"ContainerStarted","Data":"6ecc37fd4681073dd1dcc428d9daf995437306d3437f5e8499ec2781768bb8de"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.991171 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.993376 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2bhvf"] Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.994628 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj" event={"ID":"c5ad8d46-2274-448b-a6bf-e2df545679f9","Type":"ContainerStarted","Data":"a598aeedcb32f6589477dfc2654d17103e8d78b857db6ada272f39edf132058e"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.994669 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj" event={"ID":"c5ad8d46-2274-448b-a6bf-e2df545679f9","Type":"ContainerStarted","Data":"dc934d22831bf513d19dc22b61a784655d8a2f762d5c4d28485d7e2cc0e05d62"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.995496 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" event={"ID":"0b4fed59-093e-4cc0-ad3e-00825718277b","Type":"ContainerStarted","Data":"8da13d98c48b7813a7e65816eb336e513519e5e51bd2c47c5871ffaade6ac3f9"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.996783 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" event={"ID":"d50d067c-2b69-4e82-8213-a39421d81df4","Type":"ContainerStarted","Data":"163474af80950789a6cbc4e698167d1cc97a6b6b269a88639de9527602964a2c"} Oct 04 08:18:36 crc kubenswrapper[4969]: I1004 08:18:36.998182 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" event={"ID":"6379393b-2408-4b74-adf7-713fcfd448dd","Type":"ContainerStarted","Data":"02ce473d8345a9447c58bef678a43e2610fe9cf786881e3071d1702c326765c6"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.000939 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m67tq" event={"ID":"1a654328-5c34-4927-949f-0b5818da9b49","Type":"ContainerStarted","Data":"90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.004759 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" event={"ID":"baaa9767-dc9f-478a-9202-9ad67c7a9d56","Type":"ContainerStarted","Data":"468f858177f1b947af3b0c8ddf71916397348cab96ca20f64c74e2f254e32d86"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.011656 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" event={"ID":"c77eadcf-e03b-4cc7-901c-3ecd159a47bc","Type":"ContainerStarted","Data":"a2b38dc0a404774b4fdff720d2ead59f8a52ee5167542db42c37fcc9690b17b6"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.011814 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.014800 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" event={"ID":"29195139-26ff-4158-ba2c-282b1b921ede","Type":"ContainerStarted","Data":"399cabe91fe10d6fdf94a9a68bbf974f5249a2351303ee39c067de9c694a3941"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.016251 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" event={"ID":"d075b9d0-e883-4f71-a7e6-b5397d983a8b","Type":"ContainerStarted","Data":"27e22419b6ba276c2a4b04ca0f334d733ff1f4881cd8c1e49491b6e6d09c69c8"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.016303 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" event={"ID":"d075b9d0-e883-4f71-a7e6-b5397d983a8b","Type":"ContainerStarted","Data":"09dc55b1365fe479a11aa573cc32b828f61d746f8b9f4ea46420c5f76126d224"} Oct 04 08:18:37 crc kubenswrapper[4969]: W1004 08:18:37.017181 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6414b732_c979_4985_a35b_f1a45c626cc2.slice/crio-3aa7ecac8b932b048f6c42945702756350b79dd1bb50b9d698b3d68e09389419 WatchSource:0}: Error finding container 3aa7ecac8b932b048f6c42945702756350b79dd1bb50b9d698b3d68e09389419: Status 404 returned error can't find the container with id 3aa7ecac8b932b048f6c42945702756350b79dd1bb50b9d698b3d68e09389419 Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.019565 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" event={"ID":"d632f337-1d3d-42c0-b1b4-3232ed684da4","Type":"ContainerStarted","Data":"7b89fbaf74557d8fa88f26bf2be389a35907f1323a996c12cfb9159f8f1e9d1e"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.022533 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" event={"ID":"0be6fb94-7ce5-4556-88f7-0938db54d66e","Type":"ContainerStarted","Data":"e530a2fbaa241ecc3d1556cb8df67bfa795e096860ee4ba3d76864d97167e338"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.023534 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" event={"ID":"fd62df54-4f54-45fa-bcb0-92fd3ec71631","Type":"ContainerStarted","Data":"f90be0d38d935514753f0492f294a5a981b8f31f9c74eedd796dfc6c14de7752"} Oct 04 08:18:37 crc kubenswrapper[4969]: W1004 08:18:37.034105 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0a906da_f3d3_40cf_a714_2159253bb043.slice/crio-bfa614b3b46865acaad3e136a026af4269c30def36d20a5a19ce737b9f25dfab WatchSource:0}: Error finding container bfa614b3b46865acaad3e136a026af4269c30def36d20a5a19ce737b9f25dfab: Status 404 returned error can't find the container with id bfa614b3b46865acaad3e136a026af4269c30def36d20a5a19ce737b9f25dfab Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.034194 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" event={"ID":"7f61cf73-65f6-4255-8c99-4dfd6d48d80c","Type":"ContainerStarted","Data":"5dc12655eefaa97786c515da5530a463eae91587ef1f0c66f44b508abc2e00ec"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.034683 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.035033 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.535018023 +0000 UTC m=+145.289286917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.037212 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" event={"ID":"bf40de47-36a7-470e-b59b-5eea38cbc8ed","Type":"ContainerStarted","Data":"b136c69b342a46fc07cce4ab3937ccdb97179a123672f91d5d73bef632d5efd9"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.038406 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" event={"ID":"7aa45ab6-8235-4e99-af75-c60611ae22f6","Type":"ContainerStarted","Data":"a0ce95d7bafe892ffca7312db664d61faa7f59301aa0f0a7ef675bc4e8ca9fe4"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.038475 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" event={"ID":"7aa45ab6-8235-4e99-af75-c60611ae22f6","Type":"ContainerStarted","Data":"37f272dd3bb5de2188a6658fe30641d6621ab78784a6cf91ce07006d8ad445bd"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.044063 4969 generic.go:334] "Generic (PLEG): container finished" podID="2c4fb768-95d5-4587-9e36-cc0d4ce79d5b" containerID="5fed9f094c44ce4887881d23d225f393f235d6f5455f32828062f87e9b42a246" exitCode=0 Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.044162 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" event={"ID":"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b","Type":"ContainerDied","Data":"5fed9f094c44ce4887881d23d225f393f235d6f5455f32828062f87e9b42a246"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.044190 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tmqwr"] Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.044206 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" event={"ID":"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b","Type":"ContainerStarted","Data":"f275b2563002ba1cdf50536160c891845a924011a3a91a9bbd5aa7e352323312"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.052142 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" event={"ID":"846186fd-c540-464d-b85a-a67f0c8870b6","Type":"ContainerStarted","Data":"6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.052184 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" event={"ID":"846186fd-c540-464d-b85a-a67f0c8870b6","Type":"ContainerStarted","Data":"7d7c7e7c645a090f95d9cb1e9ccdd5c9dfc6d3582000563c1906ac2ef089f7e5"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.052938 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:37 crc kubenswrapper[4969]: W1004 08:18:37.068923 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f253261_82ae_49f4_9bf7_0598509f1cb2.slice/crio-2576209c979591a79e83744c5a8641a894ec96f180801a2018beb534e4ce8bc8 WatchSource:0}: Error finding container 2576209c979591a79e83744c5a8641a894ec96f180801a2018beb534e4ce8bc8: Status 404 returned error can't find the container with id 2576209c979591a79e83744c5a8641a894ec96f180801a2018beb534e4ce8bc8 Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.069030 4969 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lrtsv container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.069064 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" podUID="846186fd-c540-464d-b85a-a67f0c8870b6" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.107305 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" event={"ID":"d7581c21-cae9-4c2b-8a46-3c08bcd54180","Type":"ContainerStarted","Data":"e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.107608 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" event={"ID":"7ad46146-89bb-43b6-badb-782f113169a9","Type":"ContainerStarted","Data":"449e150fc437ee504c64edd905924ace789f07c0af92c5f54828023052c2c87f"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.107623 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.115039 4969 generic.go:334] "Generic (PLEG): container finished" podID="12536a40-76f7-448a-a0f5-0d79f111da6b" containerID="1934722c95b7c6ef0d520e5873e2028a52a433954b2b32cc8fbd4c75d056b534" exitCode=0 Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.115129 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" event={"ID":"12536a40-76f7-448a-a0f5-0d79f111da6b","Type":"ContainerDied","Data":"1934722c95b7c6ef0d520e5873e2028a52a433954b2b32cc8fbd4c75d056b534"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.120941 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" event={"ID":"f7b66b40-6c4e-41fc-ac78-e5000ce141a5","Type":"ContainerStarted","Data":"ecd75ecc8917cec45aa666595a8baa6d9f1daaabdbab0af58cf6f9d0128faa69"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.127860 4969 generic.go:334] "Generic (PLEG): container finished" podID="005c5c3d-9ede-4170-9084-b55e868faab5" containerID="4e5ebdef5bdac20b4b0b90c9f44c21cc1dbc7ca37555f4c26573fdf8edb1fbd1" exitCode=0 Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.127921 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" event={"ID":"005c5c3d-9ede-4170-9084-b55e868faab5","Type":"ContainerDied","Data":"4e5ebdef5bdac20b4b0b90c9f44c21cc1dbc7ca37555f4c26573fdf8edb1fbd1"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.138347 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.140408 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.640386065 +0000 UTC m=+145.394654879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.148129 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" event={"ID":"fcef15bf-980b-4d07-a5fb-7ab2145f69e4","Type":"ContainerStarted","Data":"8ad841a189b0d6aaef13608267a80a42d5c61eab575d8fbf7e9cd485ab6db56e"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.170270 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kxtxq" event={"ID":"f9adb985-01b8-481e-8c7a-d0e64efe0587","Type":"ContainerStarted","Data":"7bb0e365f3a1d1657258fe2aa4e9bd8fc09a9c71dbd65b5885f7fab19ff493a0"} Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.172351 4969 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfqr6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.175460 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mfqr6" podUID="b4901cb1-78b0-4aa8-9e84-22d3c1f6669c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.176767 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" Oct 04 08:18:37 crc kubenswrapper[4969]: W1004 08:18:37.197777 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7932b6aa_050e_4f49_bb0f_82ec7575d452.slice/crio-e58ad5914063b34df5375a1c6d81cf734dbe95fd233f67f1b64ab0f9a59cf2df WatchSource:0}: Error finding container e58ad5914063b34df5375a1c6d81cf734dbe95fd233f67f1b64ab0f9a59cf2df: Status 404 returned error can't find the container with id e58ad5914063b34df5375a1c6d81cf734dbe95fd233f67f1b64ab0f9a59cf2df Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.240736 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.246285 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.746272458 +0000 UTC m=+145.500541272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.341856 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.344910 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.84488308 +0000 UTC m=+145.599151904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.347655 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mglbz"] Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.367259 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99"] Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.388219 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-x9p72"] Oct 04 08:18:37 crc kubenswrapper[4969]: W1004 08:18:37.438682 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd24a7bdf_1b12_49f8_b351_896f6edf2654.slice/crio-0c6de7d84632021bdc901bab752042628ae480eb6c21cb64a28b5de783f028eb WatchSource:0}: Error finding container 0c6de7d84632021bdc901bab752042628ae480eb6c21cb64a28b5de783f028eb: Status 404 returned error can't find the container with id 0c6de7d84632021bdc901bab752042628ae480eb6c21cb64a28b5de783f028eb Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.446145 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.446474 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:37.946464625 +0000 UTC m=+145.700733439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.469833 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hbvs6" podStartSLOduration=123.469814851 podStartE2EDuration="2m3.469814851s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:37.469514243 +0000 UTC m=+145.223783057" watchObservedRunningTime="2025-10-04 08:18:37.469814851 +0000 UTC m=+145.224083665" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.547685 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.547826 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.047807494 +0000 UTC m=+145.802076308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.547988 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.548389 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.048380219 +0000 UTC m=+145.802649033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.587671 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b7p5" podStartSLOduration=124.587650984 podStartE2EDuration="2m4.587650984s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:37.579707184 +0000 UTC m=+145.333975998" watchObservedRunningTime="2025-10-04 08:18:37.587650984 +0000 UTC m=+145.341919798" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.649255 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.649410 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.14937687 +0000 UTC m=+145.903645684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.653444 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.654498 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.153978205 +0000 UTC m=+145.908247019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.654744 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" podStartSLOduration=123.654725705 podStartE2EDuration="2m3.654725705s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:37.630520598 +0000 UTC m=+145.384789412" watchObservedRunningTime="2025-10-04 08:18:37.654725705 +0000 UTC m=+145.408994519" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.717833 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-swbhw" podStartSLOduration=123.717812956 podStartE2EDuration="2m3.717812956s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:37.655468293 +0000 UTC m=+145.409737107" watchObservedRunningTime="2025-10-04 08:18:37.717812956 +0000 UTC m=+145.472081770" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.750772 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.755503 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.755859 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.255845699 +0000 UTC m=+146.010114513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.830815 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" podStartSLOduration=123.830801798 podStartE2EDuration="2m3.830801798s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:37.829010592 +0000 UTC m=+145.583279406" watchObservedRunningTime="2025-10-04 08:18:37.830801798 +0000 UTC m=+145.585070612" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.852824 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vpcfb"] Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.856638 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.856951 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.356936272 +0000 UTC m=+146.111205086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.868309 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.873332 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.921014 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vpcfb"] Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.954585 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-vbgw9" podStartSLOduration=124.954571189 podStartE2EDuration="2m4.954571189s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:37.9218555 +0000 UTC m=+145.676124314" watchObservedRunningTime="2025-10-04 08:18:37.954571189 +0000 UTC m=+145.708839993" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.961399 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.961545 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.461530544 +0000 UTC m=+146.215799358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.965815 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8kxc\" (UniqueName: \"kubernetes.io/projected/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-kube-api-access-b8kxc\") pod \"community-operators-vpcfb\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.965990 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:37 crc kubenswrapper[4969]: E1004 08:18:37.966301 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.466290372 +0000 UTC m=+146.220559186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.966446 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-utilities\") pod \"community-operators-vpcfb\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:37 crc kubenswrapper[4969]: I1004 08:18:37.966621 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-catalog-content\") pod \"community-operators-vpcfb\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.037323 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qwtnd" podStartSLOduration=125.037307302 podStartE2EDuration="2m5.037307302s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.025221229 +0000 UTC m=+145.779490043" watchObservedRunningTime="2025-10-04 08:18:38.037307302 +0000 UTC m=+145.791576116" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.037914 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b9nrp"] Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.038776 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.054008 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9nrp"] Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.060338 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wtczq" podStartSLOduration=124.06032442 podStartE2EDuration="2m4.06032442s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.05634774 +0000 UTC m=+145.810616554" watchObservedRunningTime="2025-10-04 08:18:38.06032442 +0000 UTC m=+145.814593224" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.070024 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.070231 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.570213978 +0000 UTC m=+146.324482792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.070574 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8kxc\" (UniqueName: \"kubernetes.io/projected/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-kube-api-access-b8kxc\") pod \"community-operators-vpcfb\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.070821 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.071115 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.571107069 +0000 UTC m=+146.325375883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.071976 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-utilities\") pod \"community-operators-vpcfb\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.071725 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.072560 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-utilities\") pod \"community-operators-vpcfb\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.072950 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-catalog-content\") pod \"community-operators-vpcfb\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.073459 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-catalog-content\") pod \"community-operators-vpcfb\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.122907 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8kxc\" (UniqueName: \"kubernetes.io/projected/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-kube-api-access-b8kxc\") pod \"community-operators-vpcfb\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.175280 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.175881 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-utilities\") pod \"certified-operators-b9nrp\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.176011 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-catalog-content\") pod \"certified-operators-b9nrp\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.176127 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp6g7\" (UniqueName: \"kubernetes.io/projected/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-kube-api-access-lp6g7\") pod \"certified-operators-b9nrp\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.176972 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.676957753 +0000 UTC m=+146.431226567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.189714 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6nxfz" podStartSLOduration=124.189698712 podStartE2EDuration="2m4.189698712s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.189086346 +0000 UTC m=+145.943355160" watchObservedRunningTime="2025-10-04 08:18:38.189698712 +0000 UTC m=+145.943967526" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.221733 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mfqr6" podStartSLOduration=125.221718874 podStartE2EDuration="2m5.221718874s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.216687778 +0000 UTC m=+145.970956592" watchObservedRunningTime="2025-10-04 08:18:38.221718874 +0000 UTC m=+145.975987688" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.227735 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.228289 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" event={"ID":"12536a40-76f7-448a-a0f5-0d79f111da6b","Type":"ContainerStarted","Data":"298ba66dd28bfbdbfeb6e01064512366d59aa16db149f0c2aeb48e35eb35de07"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.242457 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hvw9b"] Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.244379 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.264701 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj" event={"ID":"c5ad8d46-2274-448b-a6bf-e2df545679f9","Type":"ContainerStarted","Data":"76a2824b598fb6b16f4f140ce715ddabc17ffc84fc26ffb686fd19d0c7b6ac4b"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.267444 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" event={"ID":"6379393b-2408-4b74-adf7-713fcfd448dd","Type":"ContainerStarted","Data":"ad209d63b6eae1fca9d0b6899f4e606a2a9973e4ef8c10908f2f1270f72cae44"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.268165 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.269154 4969 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k9hcj container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.269258 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" podUID="6379393b-2408-4b74-adf7-713fcfd448dd" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.269699 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" event={"ID":"d0a906da-f3d3-40cf-a714-2159253bb043","Type":"ContainerStarted","Data":"f50e9dce1265bd029868849547362f4dd15bfd5fdccb1f64cb08a3a8c76d7448"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.269793 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" event={"ID":"d0a906da-f3d3-40cf-a714-2159253bb043","Type":"ContainerStarted","Data":"bfa614b3b46865acaad3e136a026af4269c30def36d20a5a19ce737b9f25dfab"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.270672 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4qbp6" event={"ID":"020e41ce-ceaa-4bdc-8d59-ce1dcd4f467c","Type":"ContainerStarted","Data":"549b05444ddfad2b03521383e6850ff8b6b3cbed4c684c1ba78e5f934cc0bce4"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.271647 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mglbz" event={"ID":"92b0b93f-7a67-4784-b776-c24b09b0774d","Type":"ContainerStarted","Data":"5215c8f13a1eb8e73155797f5863a1b9f14b06bb10bbfc143c54e8d5fbe889dd"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.277214 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp6g7\" (UniqueName: \"kubernetes.io/projected/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-kube-api-access-lp6g7\") pod \"certified-operators-b9nrp\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.277388 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.277563 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-utilities\") pod \"certified-operators-b9nrp\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.277637 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-catalog-content\") pod \"certified-operators-b9nrp\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.278422 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.778407515 +0000 UTC m=+146.532676329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.278946 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-utilities\") pod \"certified-operators-b9nrp\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.279223 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-catalog-content\") pod \"certified-operators-b9nrp\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.284156 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hvw9b"] Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.300217 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-x9p72" event={"ID":"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e","Type":"ContainerStarted","Data":"86895ffeea22555be025855f18e60a24da5277e0caa0a2a4661a549dc60c6680"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.308997 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp6g7\" (UniqueName: \"kubernetes.io/projected/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-kube-api-access-lp6g7\") pod \"certified-operators-b9nrp\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.309594 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" event={"ID":"7ad46146-89bb-43b6-badb-782f113169a9","Type":"ContainerStarted","Data":"cf0fdb817f3e462c38b4c231b53ddb78f297af7721f659ca5b3578befa736375"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.312216 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-m67tq" podStartSLOduration=125.312206722 podStartE2EDuration="2m5.312206722s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.311809462 +0000 UTC m=+146.066078276" watchObservedRunningTime="2025-10-04 08:18:38.312206722 +0000 UTC m=+146.066475536" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.316611 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" event={"ID":"5f151666-21ca-4ecd-ae16-6818a5780949","Type":"ContainerStarted","Data":"5a2c7f70636035beb14f50cc112d348fab430f66cd4999346e01daa60ee36ba6"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.317044 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" event={"ID":"5f151666-21ca-4ecd-ae16-6818a5780949","Type":"ContainerStarted","Data":"1e754e25a6e62387af141d97ed55d3cfc1affbc893f598b2e25c5b2eb0a877ad"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.317161 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.323312 4969 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-56p7g container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.323359 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" podUID="5f151666-21ca-4ecd-ae16-6818a5780949" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.324790 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" event={"ID":"29195139-26ff-4158-ba2c-282b1b921ede","Type":"ContainerStarted","Data":"53985fcadbe8d55b3611ba05ae14b4440661d95f4a6cc1273dcc243bd82d6faa"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.336624 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kxtxq" event={"ID":"f9adb985-01b8-481e-8c7a-d0e64efe0587","Type":"ContainerStarted","Data":"109187a785141c86c9fafec2e117a036db160c360d08d95ad8a088fb1fe1322c"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.345449 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" event={"ID":"f7b66b40-6c4e-41fc-ac78-e5000ce141a5","Type":"ContainerStarted","Data":"e092dea0acb26388ee79fcba78a0f4354aff92306c1f9c8c5928ad383ec14f80"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.347460 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" event={"ID":"bf40de47-36a7-470e-b59b-5eea38cbc8ed","Type":"ContainerStarted","Data":"bada7859f2d75a4f1f7694e39a01c3257cfb474008c3171835c34130632d0f2f"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.356506 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" podStartSLOduration=124.356491942 podStartE2EDuration="2m4.356491942s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.351968719 +0000 UTC m=+146.106237533" watchObservedRunningTime="2025-10-04 08:18:38.356491942 +0000 UTC m=+146.110760756" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.362263 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" event={"ID":"fcef15bf-980b-4d07-a5fb-7ab2145f69e4","Type":"ContainerStarted","Data":"3535fba7ca900f77e558b5424bb6c4293dab3eac9f6696c1d0d0d38d8f93c155"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.368600 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" event={"ID":"7f61cf73-65f6-4255-8c99-4dfd6d48d80c","Type":"ContainerStarted","Data":"ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.369185 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.370800 4969 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8xngh container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.31:6443/healthz\": dial tcp 10.217.0.31:6443: connect: connection refused" start-of-body= Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.370844 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" podUID="7f61cf73-65f6-4255-8c99-4dfd6d48d80c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.31:6443/healthz\": dial tcp 10.217.0.31:6443: connect: connection refused" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.371342 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" event={"ID":"e0188a25-0cb4-4e3e-8b9b-1a1a1dc47bde","Type":"ContainerStarted","Data":"24638dedd5bf63c9b5b317bad3ab903e391ca16e39c9c30d3c9c036d671186d6"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.385665 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.386769 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.886753191 +0000 UTC m=+146.641022005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.386840 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-utilities\") pod \"community-operators-hvw9b\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.386887 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdczc\" (UniqueName: \"kubernetes.io/projected/74d8fc3b-4749-4d82-92d7-df0e63457fbe-kube-api-access-bdczc\") pod \"community-operators-hvw9b\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.386904 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-catalog-content\") pod \"community-operators-hvw9b\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.386997 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.396708 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:38.89669393 +0000 UTC m=+146.650962834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.398388 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2bhvf" event={"ID":"1f253261-82ae-49f4-9bf7-0598509f1cb2","Type":"ContainerStarted","Data":"bc4b900b9cc575ce1ba7a37baa0cc1cb130d7d7393d7751e12c9097bec1fd255"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.409088 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2bhvf" event={"ID":"1f253261-82ae-49f4-9bf7-0598509f1cb2","Type":"ContainerStarted","Data":"2576209c979591a79e83744c5a8641a894ec96f180801a2018beb534e4ce8bc8"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.409117 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" event={"ID":"baaa9767-dc9f-478a-9202-9ad67c7a9d56","Type":"ContainerStarted","Data":"84a3efc10b5b366b3a35cbf8297e94efb436b9530241fa293065eb427cb69489"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.419375 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcfzj" podStartSLOduration=124.419354947 podStartE2EDuration="2m4.419354947s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.402340191 +0000 UTC m=+146.156609005" watchObservedRunningTime="2025-10-04 08:18:38.419354947 +0000 UTC m=+146.173623761" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.426334 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" event={"ID":"2c4fb768-95d5-4587-9e36-cc0d4ce79d5b","Type":"ContainerStarted","Data":"739813d61b6c78eb66d824556261e8b34cd8d9d1da9a533f95b7bfcfdc98a470"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.427038 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.455775 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" event={"ID":"6414b732-c979-4985-a35b-f1a45c626cc2","Type":"ContainerStarted","Data":"a56a0481562fc91a57d16146b7064f2fa6a46db684ed25cbbc1b20281e0f2aea"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.455969 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" event={"ID":"6414b732-c979-4985-a35b-f1a45c626cc2","Type":"ContainerStarted","Data":"3aa7ecac8b932b048f6c42945702756350b79dd1bb50b9d698b3d68e09389419"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.463882 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" event={"ID":"d24a7bdf-1b12-49f8-b351-896f6edf2654","Type":"ContainerStarted","Data":"0c6de7d84632021bdc901bab752042628ae480eb6c21cb64a28b5de783f028eb"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.467746 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" podStartSLOduration=124.467711959 podStartE2EDuration="2m4.467711959s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.440693042 +0000 UTC m=+146.194961876" watchObservedRunningTime="2025-10-04 08:18:38.467711959 +0000 UTC m=+146.221980773" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.480151 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.500165 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.500656 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.000640734 +0000 UTC m=+146.754909548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.500728 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-utilities\") pod \"community-operators-hvw9b\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.500764 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdczc\" (UniqueName: \"kubernetes.io/projected/74d8fc3b-4749-4d82-92d7-df0e63457fbe-kube-api-access-bdczc\") pod \"community-operators-hvw9b\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.505044 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-472gd"] Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.506692 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" event={"ID":"0be6fb94-7ce5-4556-88f7-0938db54d66e","Type":"ContainerStarted","Data":"8ad9b717b4bb4c3d9f9d23e28dad73e445bed88439efaac563f4ac5640669fc4"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.506822 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.509251 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-utilities\") pod \"community-operators-hvw9b\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.514790 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-catalog-content\") pod \"community-operators-hvw9b\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.514921 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.515885 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.015869936 +0000 UTC m=+146.770138750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.516311 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-472gd"] Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.516342 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" event={"ID":"5f8ffddc-8c22-4008-850c-b2d9a6738f88","Type":"ContainerStarted","Data":"82ea80d89138e9a013dff151850bac6624eec207a8e744f1f141d1416e5ba008"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.516590 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-catalog-content\") pod \"community-operators-hvw9b\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.517673 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" podStartSLOduration=125.517664881 podStartE2EDuration="2m5.517664881s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.492854339 +0000 UTC m=+146.247123153" watchObservedRunningTime="2025-10-04 08:18:38.517664881 +0000 UTC m=+146.271933695" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.525508 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" event={"ID":"fd62df54-4f54-45fa-bcb0-92fd3ec71631","Type":"ContainerStarted","Data":"3576d2d6651b7e9df2465c293c536a93543b501f15206445073d08fbb9bc0c40"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.541373 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tmqwr" event={"ID":"7932b6aa-050e-4f49-bb0f-82ec7575d452","Type":"ContainerStarted","Data":"e58ad5914063b34df5375a1c6d81cf734dbe95fd233f67f1b64ab0f9a59cf2df"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.565676 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" event={"ID":"d5aaedd0-47fa-4a7a-b61d-8776444d076b","Type":"ContainerStarted","Data":"5bdcdf2a00d4c7ec540736a7068428989b639f654b0ff64a536a5980d9158d73"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.570129 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdczc\" (UniqueName: \"kubernetes.io/projected/74d8fc3b-4749-4d82-92d7-df0e63457fbe-kube-api-access-bdczc\") pod \"community-operators-hvw9b\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.588649 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" podStartSLOduration=124.58863381 podStartE2EDuration="2m4.58863381s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.583947123 +0000 UTC m=+146.338215947" watchObservedRunningTime="2025-10-04 08:18:38.58863381 +0000 UTC m=+146.342902624" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.610002 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.611066 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" event={"ID":"0b4fed59-093e-4cc0-ad3e-00825718277b","Type":"ContainerStarted","Data":"5e4035fcbb05a6aa850c1d574d33ccb4d89a94bc6ab37ad584e180ff1afed79c"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.623499 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" event={"ID":"d50d067c-2b69-4e82-8213-a39421d81df4","Type":"ContainerStarted","Data":"ac35655ceac8bfee2dc80fa3eb478b38c8253b65edfa832a5b595e6ac09a8c06"} Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.626563 4969 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfqr6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.626589 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mfqr6" podUID="b4901cb1-78b0-4aa8-9e84-22d3c1f6669c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.635247 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.635489 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-utilities\") pod \"certified-operators-472gd\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.635535 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-catalog-content\") pod \"certified-operators-472gd\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.635553 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcftw\" (UniqueName: \"kubernetes.io/projected/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-kube-api-access-vcftw\") pod \"certified-operators-472gd\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.636333 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.136318535 +0000 UTC m=+146.890587349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.640164 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vpcfb"] Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.648902 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.662883 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-4qbp6" podStartSLOduration=6.6628671 podStartE2EDuration="6.6628671s" podCreationTimestamp="2025-10-04 08:18:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.660774927 +0000 UTC m=+146.415043741" watchObservedRunningTime="2025-10-04 08:18:38.6628671 +0000 UTC m=+146.417135914" Oct 04 08:18:38 crc kubenswrapper[4969]: W1004 08:18:38.678818 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55366ee0_2720_4e4a_83e3_c4fb9be1f6a9.slice/crio-5026eb0112bcef88d6fe02e27fde020a49d777530073c6b5eb7ff2fcd9c055f8 WatchSource:0}: Error finding container 5026eb0112bcef88d6fe02e27fde020a49d777530073c6b5eb7ff2fcd9c055f8: Status 404 returned error can't find the container with id 5026eb0112bcef88d6fe02e27fde020a49d777530073c6b5eb7ff2fcd9c055f8 Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.741151 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-utilities\") pod \"certified-operators-472gd\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.741467 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-catalog-content\") pod \"certified-operators-472gd\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.741512 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcftw\" (UniqueName: \"kubernetes.io/projected/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-kube-api-access-vcftw\") pod \"certified-operators-472gd\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.741822 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.743115 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-utilities\") pod \"certified-operators-472gd\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.744311 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-catalog-content\") pod \"certified-operators-472gd\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.745755 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.245740977 +0000 UTC m=+147.000009791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.754578 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zps22" podStartSLOduration=124.754561828 podStartE2EDuration="2m4.754561828s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.751246925 +0000 UTC m=+146.505515739" watchObservedRunningTime="2025-10-04 08:18:38.754561828 +0000 UTC m=+146.508830642" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.768373 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcftw\" (UniqueName: \"kubernetes.io/projected/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-kube-api-access-vcftw\") pod \"certified-operators-472gd\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.776759 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.783608 4969 patch_prober.go:28] interesting pod/router-default-5444994796-kxtxq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 08:18:38 crc kubenswrapper[4969]: [-]has-synced failed: reason withheld Oct 04 08:18:38 crc kubenswrapper[4969]: [+]process-running ok Oct 04 08:18:38 crc kubenswrapper[4969]: healthz check failed Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.783660 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxtxq" podUID="f9adb985-01b8-481e-8c7a-d0e64efe0587" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.844664 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.844972 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.344957524 +0000 UTC m=+147.099226338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.870963 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.889603 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-kxtxq" podStartSLOduration=124.889577312 podStartE2EDuration="2m4.889577312s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.818439149 +0000 UTC m=+146.572707963" watchObservedRunningTime="2025-10-04 08:18:38.889577312 +0000 UTC m=+146.643846126" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.924356 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" podStartSLOduration=124.924342133 podStartE2EDuration="2m4.924342133s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.890505685 +0000 UTC m=+146.644774509" watchObservedRunningTime="2025-10-04 08:18:38.924342133 +0000 UTC m=+146.678610947" Oct 04 08:18:38 crc kubenswrapper[4969]: I1004 08:18:38.948488 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:38 crc kubenswrapper[4969]: E1004 08:18:38.948788 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.448773275 +0000 UTC m=+147.203042089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.008300 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p49br" podStartSLOduration=125.008280277 podStartE2EDuration="2m5.008280277s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:38.971854034 +0000 UTC m=+146.726122848" watchObservedRunningTime="2025-10-04 08:18:39.008280277 +0000 UTC m=+146.762549091" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.009004 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-xtlw2" podStartSLOduration=125.008999505 podStartE2EDuration="2m5.008999505s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.008762789 +0000 UTC m=+146.763031603" watchObservedRunningTime="2025-10-04 08:18:39.008999505 +0000 UTC m=+146.763268319" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.050928 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.051230 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.551215962 +0000 UTC m=+147.305484766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.088757 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" podStartSLOduration=126.088740243 podStartE2EDuration="2m6.088740243s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.056609568 +0000 UTC m=+146.810878382" watchObservedRunningTime="2025-10-04 08:18:39.088740243 +0000 UTC m=+146.843009057" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.089059 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-8kcmp" podStartSLOduration=125.089052871 podStartE2EDuration="2m5.089052871s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.088982049 +0000 UTC m=+146.843250863" watchObservedRunningTime="2025-10-04 08:18:39.089052871 +0000 UTC m=+146.843321685" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.123050 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" podStartSLOduration=126.123036483 podStartE2EDuration="2m6.123036483s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.120825797 +0000 UTC m=+146.875094611" watchObservedRunningTime="2025-10-04 08:18:39.123036483 +0000 UTC m=+146.877305297" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.164693 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.165133 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.665116448 +0000 UTC m=+147.419385262 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.167838 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" podStartSLOduration=125.167820245 podStartE2EDuration="2m5.167820245s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.166515553 +0000 UTC m=+146.920784367" watchObservedRunningTime="2025-10-04 08:18:39.167820245 +0000 UTC m=+146.922089059" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.252280 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2bhvf" podStartSLOduration=7.252264241 podStartE2EDuration="7.252264241s" podCreationTimestamp="2025-10-04 08:18:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.24821235 +0000 UTC m=+147.002481164" watchObservedRunningTime="2025-10-04 08:18:39.252264241 +0000 UTC m=+147.006533055" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.252647 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9nrp"] Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.266812 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.267154 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.767127693 +0000 UTC m=+147.521396497 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.356581 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" podStartSLOduration=126.356561685 podStartE2EDuration="2m6.356561685s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.314935372 +0000 UTC m=+147.069204186" watchObservedRunningTime="2025-10-04 08:18:39.356561685 +0000 UTC m=+147.110830499" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.359533 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-45hd8" podStartSLOduration=125.359521879 podStartE2EDuration="2m5.359521879s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.351858377 +0000 UTC m=+147.106127191" watchObservedRunningTime="2025-10-04 08:18:39.359521879 +0000 UTC m=+147.113790693" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.360483 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hvw9b"] Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.367711 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.368042 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.868029003 +0000 UTC m=+147.622297807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.436256 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt9z4" podStartSLOduration=126.436238102 podStartE2EDuration="2m6.436238102s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.397613884 +0000 UTC m=+147.151882698" watchObservedRunningTime="2025-10-04 08:18:39.436238102 +0000 UTC m=+147.190506916" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.473237 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.473754 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:39.973738612 +0000 UTC m=+147.728007426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.533317 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5mtlf" podStartSLOduration=125.533296925 podStartE2EDuration="2m5.533296925s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.512317579 +0000 UTC m=+147.266586393" watchObservedRunningTime="2025-10-04 08:18:39.533296925 +0000 UTC m=+147.287565739" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.584190 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.615672 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.115654538 +0000 UTC m=+147.869923352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.654771 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tmqwr" event={"ID":"7932b6aa-050e-4f49-bb0f-82ec7575d452","Type":"ContainerStarted","Data":"1283039d5a3ab9d69706f2f85a34f6d92247f697cf3cbe552b3c9eeedc5a8bfe"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.654813 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tmqwr" event={"ID":"7932b6aa-050e-4f49-bb0f-82ec7575d452","Type":"ContainerStarted","Data":"eb5a426c4edf2bf470b38e61e7c77c37e2383d68aa5dc1288706316e2d297fb4"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.655997 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.685229 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.686232 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.186217367 +0000 UTC m=+147.940486181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.699595 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-tmqwr" podStartSLOduration=7.6995799510000005 podStartE2EDuration="7.699579951s" podCreationTimestamp="2025-10-04 08:18:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.698960986 +0000 UTC m=+147.453229800" watchObservedRunningTime="2025-10-04 08:18:39.699579951 +0000 UTC m=+147.453848765" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.706438 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9nrp" event={"ID":"46ed6532-e8c7-43c3-b80e-aeba1c5e2273","Type":"ContainerStarted","Data":"cd545504a5ec71ba8ff615bc45354d2ac0741855887563900fac792c681dd28e"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.721645 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" event={"ID":"6414b732-c979-4985-a35b-f1a45c626cc2","Type":"ContainerStarted","Data":"19d01c235779c1918c9a5a2985f97545ce232ebbfa0fc2f1002207af204987c4"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.753334 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mglbz" event={"ID":"92b0b93f-7a67-4784-b776-c24b09b0774d","Type":"ContainerStarted","Data":"a2d8090f6afab0c9a0943b07eba4501f56bb46b7569a818b03f63df98021a9eb"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.753682 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.764768 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ddzpn" podStartSLOduration=125.764753145 podStartE2EDuration="2m5.764753145s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.763785081 +0000 UTC m=+147.518053885" watchObservedRunningTime="2025-10-04 08:18:39.764753145 +0000 UTC m=+147.519021959" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.767556 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" event={"ID":"7ad46146-89bb-43b6-badb-782f113169a9","Type":"ContainerStarted","Data":"6a2089c8ae0cc6282ffc93e2591971d40aa0eff351838533b7052137a5dd0a1d"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.781244 4969 patch_prober.go:28] interesting pod/console-operator-58897d9998-mglbz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.781291 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-mglbz" podUID="92b0b93f-7a67-4784-b776-c24b09b0774d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.785350 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" event={"ID":"d24a7bdf-1b12-49f8-b351-896f6edf2654","Type":"ContainerStarted","Data":"f6ac9af222e2a6dfd87defa290f80938642bb5a03aa859f79347a08ae512b4df"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.786509 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.787536 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.287522405 +0000 UTC m=+148.041791219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.801168 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvw9b" event={"ID":"74d8fc3b-4749-4d82-92d7-df0e63457fbe","Type":"ContainerStarted","Data":"a16e838189f8c416132621aeaad5e9f8b2c7b70df6f369bb6ce201935d53675f"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.801628 4969 patch_prober.go:28] interesting pod/router-default-5444994796-kxtxq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 08:18:39 crc kubenswrapper[4969]: [-]has-synced failed: reason withheld Oct 04 08:18:39 crc kubenswrapper[4969]: [+]process-running ok Oct 04 08:18:39 crc kubenswrapper[4969]: healthz check failed Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.801657 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxtxq" podUID="f9adb985-01b8-481e-8c7a-d0e64efe0587" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.845920 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" event={"ID":"005c5c3d-9ede-4170-9084-b55e868faab5","Type":"ContainerStarted","Data":"3822923c1d0c9f678ec6dfb779e287db22f35a3fff4f516b8e5d1942f723d123"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.845964 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" event={"ID":"005c5c3d-9ede-4170-9084-b55e868faab5","Type":"ContainerStarted","Data":"d1e963edc3ffaf92aa9ddfe73437ecb204609de21b097ae89ce4aa3c4e0be8e1"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.857214 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-mglbz" podStartSLOduration=126.857199042 podStartE2EDuration="2m6.857199042s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.811894477 +0000 UTC m=+147.566163291" watchObservedRunningTime="2025-10-04 08:18:39.857199042 +0000 UTC m=+147.611467856" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.858850 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rlrqv"] Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.859812 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.876214 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.878690 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-472gd"] Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.885777 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2ksr9" event={"ID":"0be6fb94-7ce5-4556-88f7-0938db54d66e","Type":"ContainerStarted","Data":"a0bd11dbbf661ea564a6eb75ca23c5430266577e8d5fa19a85675ec42509e31e"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.887213 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.887458 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.387412519 +0000 UTC m=+148.141681333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.887518 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-catalog-content\") pod \"redhat-marketplace-rlrqv\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.887814 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-utilities\") pod \"redhat-marketplace-rlrqv\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.887937 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.887970 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhgqn\" (UniqueName: \"kubernetes.io/projected/03b29be6-fac6-418e-97f1-a9cdad90fbf0-kube-api-access-rhgqn\") pod \"redhat-marketplace-rlrqv\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:39 crc kubenswrapper[4969]: E1004 08:18:39.888961 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.388939978 +0000 UTC m=+148.143208872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.889642 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mtn99" podStartSLOduration=125.889626924 podStartE2EDuration="2m5.889626924s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.876608649 +0000 UTC m=+147.630877463" watchObservedRunningTime="2025-10-04 08:18:39.889626924 +0000 UTC m=+147.643895738" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.899500 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlrqv"] Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.904228 4969 generic.go:334] "Generic (PLEG): container finished" podID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerID="82be18e59c05e559f6ddb241d8530a4817aea0513a0d0287256b93d7dd615780" exitCode=0 Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.904304 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpcfb" event={"ID":"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9","Type":"ContainerDied","Data":"82be18e59c05e559f6ddb241d8530a4817aea0513a0d0287256b93d7dd615780"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.904333 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpcfb" event={"ID":"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9","Type":"ContainerStarted","Data":"5026eb0112bcef88d6fe02e27fde020a49d777530073c6b5eb7ff2fcd9c055f8"} Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.915332 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.915798 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h8gcd" podStartSLOduration=125.91577916 podStartE2EDuration="2m5.91577916s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:39.914870897 +0000 UTC m=+147.669139711" watchObservedRunningTime="2025-10-04 08:18:39.91577916 +0000 UTC m=+147.670047994" Oct 04 08:18:39 crc kubenswrapper[4969]: I1004 08:18:39.972073 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2lsw" event={"ID":"5f8ffddc-8c22-4008-850c-b2d9a6738f88","Type":"ContainerStarted","Data":"387c343cd8f232ced616bbeb6b7ed191f23d2920330d2e3d83d474e88dffb0b3"} Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.008381 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.008844 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhgqn\" (UniqueName: \"kubernetes.io/projected/03b29be6-fac6-418e-97f1-a9cdad90fbf0-kube-api-access-rhgqn\") pod \"redhat-marketplace-rlrqv\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.008919 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-catalog-content\") pod \"redhat-marketplace-rlrqv\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.009088 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-utilities\") pod \"redhat-marketplace-rlrqv\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.010150 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.510134005 +0000 UTC m=+148.264402889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.011189 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-catalog-content\") pod \"redhat-marketplace-rlrqv\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.012131 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-utilities\") pod \"redhat-marketplace-rlrqv\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.034959 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.035312 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.040350 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" event={"ID":"fcef15bf-980b-4d07-a5fb-7ab2145f69e4","Type":"ContainerStarted","Data":"bb258eedc65b34d4a0e7346d555384d85f9d1e3d56ddd34a26eff0927ba82253"} Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.073371 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.077340 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k9hcj" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.078170 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhgqn\" (UniqueName: \"kubernetes.io/projected/03b29be6-fac6-418e-97f1-a9cdad90fbf0-kube-api-access-rhgqn\") pod \"redhat-marketplace-rlrqv\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.096457 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" podStartSLOduration=127.096421597 podStartE2EDuration="2m7.096421597s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:40.09610808 +0000 UTC m=+147.850376894" watchObservedRunningTime="2025-10-04 08:18:40.096421597 +0000 UTC m=+147.850690411" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.113723 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.124242 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.624225494 +0000 UTC m=+148.378494308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.181903 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.181950 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.187831 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.201291 4969 patch_prober.go:28] interesting pod/apiserver-76f77b778f-t4b7l container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.201343 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" podUID="005c5c3d-9ede-4170-9084-b55e868faab5" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.223039 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.223251 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.723233905 +0000 UTC m=+148.477502719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.223531 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.223873 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.723862691 +0000 UTC m=+148.478131495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.304382 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xtfkc" podStartSLOduration=126.304367588 podStartE2EDuration="2m6.304367588s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:40.268342936 +0000 UTC m=+148.022611750" watchObservedRunningTime="2025-10-04 08:18:40.304367588 +0000 UTC m=+148.058636402" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.304723 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zfxkq"] Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.319999 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.326280 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.326518 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.826504343 +0000 UTC m=+148.580773157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.326608 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.326901 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.826894043 +0000 UTC m=+148.581162857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.368179 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rc9xt" podStartSLOduration=126.368163558 podStartE2EDuration="2m6.368163558s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:40.35471401 +0000 UTC m=+148.108982824" watchObservedRunningTime="2025-10-04 08:18:40.368163558 +0000 UTC m=+148.122432372" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.370580 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfxkq"] Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.429922 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.430124 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-utilities\") pod \"redhat-marketplace-zfxkq\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.430220 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-catalog-content\") pod \"redhat-marketplace-zfxkq\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.430248 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96b4z\" (UniqueName: \"kubernetes.io/projected/072c05bc-8cfa-40d0-ba2b-41af13739f3f-kube-api-access-96b4z\") pod \"redhat-marketplace-zfxkq\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.430333 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:40.930319245 +0000 UTC m=+148.684588059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.512257 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" podStartSLOduration=126.512243989 podStartE2EDuration="2m6.512243989s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:40.439653209 +0000 UTC m=+148.193922023" watchObservedRunningTime="2025-10-04 08:18:40.512243989 +0000 UTC m=+148.266512803" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.536625 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-utilities\") pod \"redhat-marketplace-zfxkq\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.536913 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.536949 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-catalog-content\") pod \"redhat-marketplace-zfxkq\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.536982 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96b4z\" (UniqueName: \"kubernetes.io/projected/072c05bc-8cfa-40d0-ba2b-41af13739f3f-kube-api-access-96b4z\") pod \"redhat-marketplace-zfxkq\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.537779 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.037761928 +0000 UTC m=+148.792030742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.542853 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-utilities\") pod \"redhat-marketplace-zfxkq\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.544069 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-catalog-content\") pod \"redhat-marketplace-zfxkq\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.584252 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96b4z\" (UniqueName: \"kubernetes.io/projected/072c05bc-8cfa-40d0-ba2b-41af13739f3f-kube-api-access-96b4z\") pod \"redhat-marketplace-zfxkq\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.638051 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.638444 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.13841291 +0000 UTC m=+148.892681724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.704032 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.746110 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.746507 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.246490279 +0000 UTC m=+149.000759093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.783543 4969 patch_prober.go:28] interesting pod/router-default-5444994796-kxtxq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 08:18:40 crc kubenswrapper[4969]: [-]has-synced failed: reason withheld Oct 04 08:18:40 crc kubenswrapper[4969]: [+]process-running ok Oct 04 08:18:40 crc kubenswrapper[4969]: healthz check failed Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.783592 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxtxq" podUID="f9adb985-01b8-481e-8c7a-d0e64efe0587" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.846855 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.847228 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.347213823 +0000 UTC m=+149.101482637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.884906 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlrqv"] Oct 04 08:18:40 crc kubenswrapper[4969]: I1004 08:18:40.952078 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:40 crc kubenswrapper[4969]: E1004 08:18:40.952401 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.452388949 +0000 UTC m=+149.206657763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.049149 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mxq79"] Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.049577 4969 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8xngh container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.31:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.049675 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" podUID="7f61cf73-65f6-4255-8c99-4dfd6d48d80c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.31:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.050370 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.056559 4969 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-56p7g container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.056642 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" podUID="5f151666-21ca-4ecd-ae16-6818a5780949" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.056855 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.057061 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.057112 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.057139 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.057175 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.059183 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.559164064 +0000 UTC m=+149.313432878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.060613 4969 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-nhfgd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.060658 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" podUID="2c4fb768-95d5-4587-9e36-cc0d4ce79d5b" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.064889 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.066881 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.079551 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.088182 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxq79"] Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.094816 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlrqv" event={"ID":"03b29be6-fac6-418e-97f1-a9cdad90fbf0","Type":"ContainerStarted","Data":"ef2cbb055512836775d751923f304b924bfcc4b3ea4fc258ae6456b4f6f9f5a6"} Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.099230 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.116025 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.152636 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-x9p72" event={"ID":"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e","Type":"ContainerStarted","Data":"97eb543ec62b97bdba2d5ed86c87a57fb20f991e0a2b18a9deff0413cfdc4500"} Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.160107 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzzwv\" (UniqueName: \"kubernetes.io/projected/f0a988d3-5ee9-4833-97c8-05796123ff9a-kube-api-access-vzzwv\") pod \"redhat-operators-mxq79\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.160159 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-catalog-content\") pod \"redhat-operators-mxq79\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.160188 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.160212 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-utilities\") pod \"redhat-operators-mxq79\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.160471 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.660460274 +0000 UTC m=+149.414729088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.178775 4969 generic.go:334] "Generic (PLEG): container finished" podID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerID="532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8" exitCode=0 Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.178839 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472gd" event={"ID":"b7a10e08-6a22-4d0a-af3a-b8bc8c234830","Type":"ContainerDied","Data":"532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8"} Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.178863 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472gd" event={"ID":"b7a10e08-6a22-4d0a-af3a-b8bc8c234830","Type":"ContainerStarted","Data":"bf669b959e3f52c02f28f4ea079b2f01b7c387e2d78148182a74cbc70e60a7f5"} Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.192801 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.229765 4969 generic.go:334] "Generic (PLEG): container finished" podID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerID="5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d" exitCode=0 Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.229840 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvw9b" event={"ID":"74d8fc3b-4749-4d82-92d7-df0e63457fbe","Type":"ContainerDied","Data":"5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d"} Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.254609 4969 generic.go:334] "Generic (PLEG): container finished" podID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerID="6389c63785aca13aac72855d1369bd2d3a94697d57c6138bb11ba42dbb5a1248" exitCode=0 Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.255713 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9nrp" event={"ID":"46ed6532-e8c7-43c3-b80e-aeba1c5e2273","Type":"ContainerDied","Data":"6389c63785aca13aac72855d1369bd2d3a94697d57c6138bb11ba42dbb5a1248"} Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.266899 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.267456 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-catalog-content\") pod \"redhat-operators-mxq79\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.268168 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.768143883 +0000 UTC m=+149.522412697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.268227 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-utilities\") pod \"redhat-operators-mxq79\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.268340 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzzwv\" (UniqueName: \"kubernetes.io/projected/f0a988d3-5ee9-4833-97c8-05796123ff9a-kube-api-access-vzzwv\") pod \"redhat-operators-mxq79\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.268683 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-catalog-content\") pod \"redhat-operators-mxq79\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.269032 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-utilities\") pod \"redhat-operators-mxq79\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.270658 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.281489 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xcxzt" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.281676 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.323124 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzzwv\" (UniqueName: \"kubernetes.io/projected/f0a988d3-5ee9-4833-97c8-05796123ff9a-kube-api-access-vzzwv\") pod \"redhat-operators-mxq79\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.371239 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.382393 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.882379325 +0000 UTC m=+149.636648139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.417708 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.485985 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.486497 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:41.986481915 +0000 UTC m=+149.740750729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.488586 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-mglbz" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.494360 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-db8bm"] Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.495305 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.541213 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nhfgd" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.561221 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-db8bm"] Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.586052 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfxkq"] Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.588093 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-utilities\") pod \"redhat-operators-db8bm\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.588129 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-catalog-content\") pod \"redhat-operators-db8bm\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.588158 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.588193 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpn9q\" (UniqueName: \"kubernetes.io/projected/724eefc4-1c68-4593-8aaf-0144f80e3784-kube-api-access-kpn9q\") pod \"redhat-operators-db8bm\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.588488 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.08847632 +0000 UTC m=+149.842745134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.689769 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.689969 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.189943303 +0000 UTC m=+149.944212117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.690085 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-utilities\") pod \"redhat-operators-db8bm\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.690113 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-catalog-content\") pod \"redhat-operators-db8bm\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.690150 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.690213 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpn9q\" (UniqueName: \"kubernetes.io/projected/724eefc4-1c68-4593-8aaf-0144f80e3784-kube-api-access-kpn9q\") pod \"redhat-operators-db8bm\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.690891 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-utilities\") pod \"redhat-operators-db8bm\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.691131 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-catalog-content\") pod \"redhat-operators-db8bm\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.691366 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.191355749 +0000 UTC m=+149.945624573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.729572 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpn9q\" (UniqueName: \"kubernetes.io/projected/724eefc4-1c68-4593-8aaf-0144f80e3784-kube-api-access-kpn9q\") pod \"redhat-operators-db8bm\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.784952 4969 patch_prober.go:28] interesting pod/router-default-5444994796-kxtxq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 08:18:41 crc kubenswrapper[4969]: [-]has-synced failed: reason withheld Oct 04 08:18:41 crc kubenswrapper[4969]: [+]process-running ok Oct 04 08:18:41 crc kubenswrapper[4969]: healthz check failed Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.785014 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxtxq" podUID="f9adb985-01b8-481e-8c7a-d0e64efe0587" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.793942 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.794238 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.294222216 +0000 UTC m=+150.048491030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.848798 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.884618 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.885252 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.890755 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.890956 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.896210 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:41 crc kubenswrapper[4969]: E1004 08:18:41.896485 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.396472769 +0000 UTC m=+150.150741583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:41 crc kubenswrapper[4969]: I1004 08:18:41.903565 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.022071 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.022830 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.022942 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.023111 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.523089283 +0000 UTC m=+150.277358097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.124304 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.124389 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.124449 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.124792 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.124900 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.624889754 +0000 UTC m=+150.379158568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.152641 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxq79"] Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.160120 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.205198 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.224963 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.225287 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.725272579 +0000 UTC m=+150.479541383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.325932 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.326199 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.826186658 +0000 UTC m=+150.580455472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.328127 4969 generic.go:334] "Generic (PLEG): container finished" podID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerID="be80e8d8059750c4949b13ea155aa7a6845fd994b5bd786b661109315328fabc" exitCode=0 Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.328192 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlrqv" event={"ID":"03b29be6-fac6-418e-97f1-a9cdad90fbf0","Type":"ContainerDied","Data":"be80e8d8059750c4949b13ea155aa7a6845fd994b5bd786b661109315328fabc"} Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.353760 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ead41343e9dcaddcbcd2b52b9e90fff236fbf860bd150814d3a48588403b0327"} Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.355701 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cbaae936f5e879a0e8d5d75752210b44cde777730a42557db46460cff5d48c3c"} Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.361951 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxq79" event={"ID":"f0a988d3-5ee9-4833-97c8-05796123ff9a","Type":"ContainerStarted","Data":"729be080ffac341b25560a4365ab215d476ac6a7835a379d7ee60d06eb438f60"} Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.365208 4969 generic.go:334] "Generic (PLEG): container finished" podID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerID="83aaabc54fac748acc33f873266b06f0555915c22b5b6f1ed76cee2ee3e15f64" exitCode=0 Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.366318 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfxkq" event={"ID":"072c05bc-8cfa-40d0-ba2b-41af13739f3f","Type":"ContainerDied","Data":"83aaabc54fac748acc33f873266b06f0555915c22b5b6f1ed76cee2ee3e15f64"} Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.366350 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfxkq" event={"ID":"072c05bc-8cfa-40d0-ba2b-41af13739f3f","Type":"ContainerStarted","Data":"0c7e57e11ba180d9da2f049379d8a12fa3b01038fe90e8e8cc3e40bde15eface"} Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.427234 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.427439 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.927402745 +0000 UTC m=+150.681671549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.427899 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.429619 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:42.92960369 +0000 UTC m=+150.683872494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.529615 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.529736 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.029715339 +0000 UTC m=+150.783984163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.529992 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.530374 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.030364255 +0000 UTC m=+150.784633069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.565516 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-db8bm"] Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.565824 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.630576 4969 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.630793 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.630909 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.130888465 +0000 UTC m=+150.885157279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.631263 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.631592 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.131585622 +0000 UTC m=+150.885854436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.732169 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.732309 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.232285006 +0000 UTC m=+150.986553820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.732486 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.732794 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.232787449 +0000 UTC m=+150.987056253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: W1004 08:18:42.740144 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc4c27be4_3761_4dfe_9f29_22371c4e62a3.slice/crio-b12ad062e6d54b215ba2c1e3c3b712ce4b1df05f601a14915917f59415d42c10 WatchSource:0}: Error finding container b12ad062e6d54b215ba2c1e3c3b712ce4b1df05f601a14915917f59415d42c10: Status 404 returned error can't find the container with id b12ad062e6d54b215ba2c1e3c3b712ce4b1df05f601a14915917f59415d42c10 Oct 04 08:18:42 crc kubenswrapper[4969]: W1004 08:18:42.743711 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod724eefc4_1c68_4593_8aaf_0144f80e3784.slice/crio-2ea5fdb2c2afbb1970f7d6213bff7b53734f071b6b627015cd3c8c79a99018a3 WatchSource:0}: Error finding container 2ea5fdb2c2afbb1970f7d6213bff7b53734f071b6b627015cd3c8c79a99018a3: Status 404 returned error can't find the container with id 2ea5fdb2c2afbb1970f7d6213bff7b53734f071b6b627015cd3c8c79a99018a3 Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.783004 4969 patch_prober.go:28] interesting pod/router-default-5444994796-kxtxq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 08:18:42 crc kubenswrapper[4969]: [-]has-synced failed: reason withheld Oct 04 08:18:42 crc kubenswrapper[4969]: [+]process-running ok Oct 04 08:18:42 crc kubenswrapper[4969]: healthz check failed Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.783353 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxtxq" podUID="f9adb985-01b8-481e-8c7a-d0e64efe0587" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.842819 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.843161 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.343144094 +0000 UTC m=+151.097412908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:42 crc kubenswrapper[4969]: I1004 08:18:42.944903 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:42 crc kubenswrapper[4969]: E1004 08:18:42.945250 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.445234892 +0000 UTC m=+151.199503706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.047010 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:43 crc kubenswrapper[4969]: E1004 08:18:43.047227 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.547182357 +0000 UTC m=+151.301451171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.047364 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:43 crc kubenswrapper[4969]: E1004 08:18:43.047801 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.547791233 +0000 UTC m=+151.302060047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.149667 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:43 crc kubenswrapper[4969]: E1004 08:18:43.149994 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.649980504 +0000 UTC m=+151.404249308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.252188 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:43 crc kubenswrapper[4969]: E1004 08:18:43.252571 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 08:18:43.752556535 +0000 UTC m=+151.506825349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cbwg8" (UID: "07fc39da-f348-4892-a8ab-df8200d861af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.267051 4969 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-04T08:18:42.630600018Z","Handler":null,"Name":""} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.272194 4969 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.272232 4969 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.353663 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.356826 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.382348 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a6158dbd3d3dd591a64c3fc8f22a024ae283b18ca97d059e1fd4e0dedcc3ef20"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.383046 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.388995 4969 generic.go:334] "Generic (PLEG): container finished" podID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerID="3f225ef56d508ec4094f2830bc4ae2b8c9a88d142c7da2ea693f46e09fd3aed3" exitCode=0 Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.389049 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db8bm" event={"ID":"724eefc4-1c68-4593-8aaf-0144f80e3784","Type":"ContainerDied","Data":"3f225ef56d508ec4094f2830bc4ae2b8c9a88d142c7da2ea693f46e09fd3aed3"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.389071 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db8bm" event={"ID":"724eefc4-1c68-4593-8aaf-0144f80e3784","Type":"ContainerStarted","Data":"2ea5fdb2c2afbb1970f7d6213bff7b53734f071b6b627015cd3c8c79a99018a3"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.391924 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c4c27be4-3761-4dfe-9f29-22371c4e62a3","Type":"ContainerStarted","Data":"c4520af2f7e0df3e5c17d8e8d15b03eb1250eb0c4d90a6f20e9c7fcbbb4b6936"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.391952 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c4c27be4-3761-4dfe-9f29-22371c4e62a3","Type":"ContainerStarted","Data":"b12ad062e6d54b215ba2c1e3c3b712ce4b1df05f601a14915917f59415d42c10"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.402569 4969 generic.go:334] "Generic (PLEG): container finished" podID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerID="f52a5bbb0c745547ca030f2fcbd2a7057b9eae68db08f6c3ecc1fb44ffa2031e" exitCode=0 Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.402727 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxq79" event={"ID":"f0a988d3-5ee9-4833-97c8-05796123ff9a","Type":"ContainerDied","Data":"f52a5bbb0c745547ca030f2fcbd2a7057b9eae68db08f6c3ecc1fb44ffa2031e"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.416178 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2415f859e588602de8f9f78b2099f46d0517da80f9dfa513a562cc96b6d4899b"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.416218 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"48d94308ec860ea398d383c3d08de50a4ae58c98b8eb259b32d782fa02d3e5ec"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.427122 4969 generic.go:334] "Generic (PLEG): container finished" podID="29195139-26ff-4158-ba2c-282b1b921ede" containerID="53985fcadbe8d55b3611ba05ae14b4440661d95f4a6cc1273dcc243bd82d6faa" exitCode=0 Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.427231 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" event={"ID":"29195139-26ff-4158-ba2c-282b1b921ede","Type":"ContainerDied","Data":"53985fcadbe8d55b3611ba05ae14b4440661d95f4a6cc1273dcc243bd82d6faa"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.430090 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.430075284 podStartE2EDuration="2.430075284s" podCreationTimestamp="2025-10-04 08:18:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:43.425224412 +0000 UTC m=+151.179493226" watchObservedRunningTime="2025-10-04 08:18:43.430075284 +0000 UTC m=+151.184344098" Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.431508 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d5a5656efe793e62ac0de51e5c561f261dd6285e5b58277c8a0f521760b7511a"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.440781 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-x9p72" event={"ID":"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e","Type":"ContainerStarted","Data":"3a20cadb142a0cb4a11639b00cfac11ddbae9de50a94ade3f5d1a0b22af7ee0e"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.440821 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-x9p72" event={"ID":"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e","Type":"ContainerStarted","Data":"5c5b20a6404ab57dd834e092ed6ff607054a1b554934ebd43cf563d6b54e9e7c"} Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.458059 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.462541 4969 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.462576 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.528725 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cbwg8\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.610812 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.780257 4969 patch_prober.go:28] interesting pod/router-default-5444994796-kxtxq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 08:18:43 crc kubenswrapper[4969]: [-]has-synced failed: reason withheld Oct 04 08:18:43 crc kubenswrapper[4969]: [+]process-running ok Oct 04 08:18:43 crc kubenswrapper[4969]: healthz check failed Oct 04 08:18:43 crc kubenswrapper[4969]: I1004 08:18:43.780513 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxtxq" podUID="f9adb985-01b8-481e-8c7a-d0e64efe0587" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.272139 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cbwg8"] Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.325187 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.474623 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-x9p72" event={"ID":"d8eb4ca9-7b43-45f4-ae87-ceffc3f6d57e","Type":"ContainerStarted","Data":"a901a73890cabb8779b74ba726455496c61cb691c313181c1880c0344215c7bf"} Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.477341 4969 generic.go:334] "Generic (PLEG): container finished" podID="c4c27be4-3761-4dfe-9f29-22371c4e62a3" containerID="c4520af2f7e0df3e5c17d8e8d15b03eb1250eb0c4d90a6f20e9c7fcbbb4b6936" exitCode=0 Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.477486 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c4c27be4-3761-4dfe-9f29-22371c4e62a3","Type":"ContainerDied","Data":"c4520af2f7e0df3e5c17d8e8d15b03eb1250eb0c4d90a6f20e9c7fcbbb4b6936"} Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.481790 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" event={"ID":"07fc39da-f348-4892-a8ab-df8200d861af","Type":"ContainerStarted","Data":"0dd6a1d7c0c68e132d7d5846049143fbb1150bcba6149ad1932dfa8de677bbe2"} Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.498909 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-x9p72" podStartSLOduration=12.49889102 podStartE2EDuration="12.49889102s" podCreationTimestamp="2025-10-04 08:18:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:44.497903195 +0000 UTC m=+152.252172009" watchObservedRunningTime="2025-10-04 08:18:44.49889102 +0000 UTC m=+152.253159834" Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.776701 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.780154 4969 patch_prober.go:28] interesting pod/router-default-5444994796-kxtxq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 08:18:44 crc kubenswrapper[4969]: [-]has-synced failed: reason withheld Oct 04 08:18:44 crc kubenswrapper[4969]: [+]process-running ok Oct 04 08:18:44 crc kubenswrapper[4969]: healthz check failed Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.780205 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxtxq" podUID="f9adb985-01b8-481e-8c7a-d0e64efe0587" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.910691 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bht6\" (UniqueName: \"kubernetes.io/projected/29195139-26ff-4158-ba2c-282b1b921ede-kube-api-access-6bht6\") pod \"29195139-26ff-4158-ba2c-282b1b921ede\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.910970 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29195139-26ff-4158-ba2c-282b1b921ede-config-volume\") pod \"29195139-26ff-4158-ba2c-282b1b921ede\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.911053 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29195139-26ff-4158-ba2c-282b1b921ede-secret-volume\") pod \"29195139-26ff-4158-ba2c-282b1b921ede\" (UID: \"29195139-26ff-4158-ba2c-282b1b921ede\") " Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.920839 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29195139-26ff-4158-ba2c-282b1b921ede-config-volume" (OuterVolumeSpecName: "config-volume") pod "29195139-26ff-4158-ba2c-282b1b921ede" (UID: "29195139-26ff-4158-ba2c-282b1b921ede"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.921687 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29195139-26ff-4158-ba2c-282b1b921ede-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "29195139-26ff-4158-ba2c-282b1b921ede" (UID: "29195139-26ff-4158-ba2c-282b1b921ede"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:18:44 crc kubenswrapper[4969]: I1004 08:18:44.921796 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29195139-26ff-4158-ba2c-282b1b921ede-kube-api-access-6bht6" (OuterVolumeSpecName: "kube-api-access-6bht6") pod "29195139-26ff-4158-ba2c-282b1b921ede" (UID: "29195139-26ff-4158-ba2c-282b1b921ede"). InnerVolumeSpecName "kube-api-access-6bht6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.013509 4969 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29195139-26ff-4158-ba2c-282b1b921ede-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.013541 4969 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29195139-26ff-4158-ba2c-282b1b921ede-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.013551 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bht6\" (UniqueName: \"kubernetes.io/projected/29195139-26ff-4158-ba2c-282b1b921ede-kube-api-access-6bht6\") on node \"crc\" DevicePath \"\"" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.050622 4969 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfqr6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.050683 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mfqr6" podUID="b4901cb1-78b0-4aa8-9e84-22d3c1f6669c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.051056 4969 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfqr6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.051077 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mfqr6" podUID="b4901cb1-78b0-4aa8-9e84-22d3c1f6669c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.067101 4969 patch_prober.go:28] interesting pod/console-f9d7485db-m67tq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.067150 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m67tq" podUID="1a654328-5c34-4927-949f-0b5818da9b49" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.067232 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.067946 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.067973 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.191164 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.196258 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-t4b7l" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.490616 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" event={"ID":"07fc39da-f348-4892-a8ab-df8200d861af","Type":"ContainerStarted","Data":"43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743"} Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.490812 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.495815 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.496238 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x" event={"ID":"29195139-26ff-4158-ba2c-282b1b921ede","Type":"ContainerDied","Data":"399cabe91fe10d6fdf94a9a68bbf974f5249a2351303ee39c067de9c694a3941"} Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.496255 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="399cabe91fe10d6fdf94a9a68bbf974f5249a2351303ee39c067de9c694a3941" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.527305 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" podStartSLOduration=131.527287373 podStartE2EDuration="2m11.527287373s" podCreationTimestamp="2025-10-04 08:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:45.51477273 +0000 UTC m=+153.269041554" watchObservedRunningTime="2025-10-04 08:18:45.527287373 +0000 UTC m=+153.281556177" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.586524 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.775906 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.779034 4969 patch_prober.go:28] interesting pod/router-default-5444994796-kxtxq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 08:18:45 crc kubenswrapper[4969]: [-]has-synced failed: reason withheld Oct 04 08:18:45 crc kubenswrapper[4969]: [+]process-running ok Oct 04 08:18:45 crc kubenswrapper[4969]: healthz check failed Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.779088 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxtxq" podUID="f9adb985-01b8-481e-8c7a-d0e64efe0587" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.786227 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.936544 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kubelet-dir\") pod \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\" (UID: \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\") " Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.936668 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kube-api-access\") pod \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\" (UID: \"c4c27be4-3761-4dfe-9f29-22371c4e62a3\") " Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.937906 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c4c27be4-3761-4dfe-9f29-22371c4e62a3" (UID: "c4c27be4-3761-4dfe-9f29-22371c4e62a3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.940684 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c4c27be4-3761-4dfe-9f29-22371c4e62a3" (UID: "c4c27be4-3761-4dfe-9f29-22371c4e62a3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.964412 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.967637 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:18:45 crc kubenswrapper[4969]: I1004 08:18:45.989526 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-56p7g" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.038029 4969 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.038475 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4c27be4-3761-4dfe-9f29-22371c4e62a3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.116873 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 08:18:46 crc kubenswrapper[4969]: E1004 08:18:46.117534 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29195139-26ff-4158-ba2c-282b1b921ede" containerName="collect-profiles" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.117638 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="29195139-26ff-4158-ba2c-282b1b921ede" containerName="collect-profiles" Oct 04 08:18:46 crc kubenswrapper[4969]: E1004 08:18:46.117669 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c27be4-3761-4dfe-9f29-22371c4e62a3" containerName="pruner" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.117677 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c27be4-3761-4dfe-9f29-22371c4e62a3" containerName="pruner" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.117844 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4c27be4-3761-4dfe-9f29-22371c4e62a3" containerName="pruner" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.117862 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="29195139-26ff-4158-ba2c-282b1b921ede" containerName="collect-profiles" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.118569 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.120771 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.121090 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.135148 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.240858 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2373497a-f79e-4f48-8792-4855301ff00d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2373497a-f79e-4f48-8792-4855301ff00d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.240932 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2373497a-f79e-4f48-8792-4855301ff00d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2373497a-f79e-4f48-8792-4855301ff00d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.344638 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2373497a-f79e-4f48-8792-4855301ff00d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2373497a-f79e-4f48-8792-4855301ff00d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.344734 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2373497a-f79e-4f48-8792-4855301ff00d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2373497a-f79e-4f48-8792-4855301ff00d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.344808 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2373497a-f79e-4f48-8792-4855301ff00d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2373497a-f79e-4f48-8792-4855301ff00d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.370666 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2373497a-f79e-4f48-8792-4855301ff00d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2373497a-f79e-4f48-8792-4855301ff00d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.448309 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.552114 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.552332 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c4c27be4-3761-4dfe-9f29-22371c4e62a3","Type":"ContainerDied","Data":"b12ad062e6d54b215ba2c1e3c3b712ce4b1df05f601a14915917f59415d42c10"} Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.552376 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b12ad062e6d54b215ba2c1e3c3b712ce4b1df05f601a14915917f59415d42c10" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.779392 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.782197 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-kxtxq" Oct 04 08:18:46 crc kubenswrapper[4969]: I1004 08:18:46.985332 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 08:18:47 crc kubenswrapper[4969]: I1004 08:18:47.575985 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2373497a-f79e-4f48-8792-4855301ff00d","Type":"ContainerStarted","Data":"e52d9a631dbba75e65e43a9b55ec3990be505e884678baa5561dc3a0b2977cd5"} Oct 04 08:18:48 crc kubenswrapper[4969]: I1004 08:18:48.003238 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-tmqwr" Oct 04 08:18:49 crc kubenswrapper[4969]: I1004 08:18:49.608964 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2373497a-f79e-4f48-8792-4855301ff00d","Type":"ContainerStarted","Data":"7a2f1fe7792bcacc4efec54cfe9c2eb184d621852cc7e78c6a2a7050fd818e10"} Oct 04 08:18:49 crc kubenswrapper[4969]: I1004 08:18:49.630575 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.630557497 podStartE2EDuration="3.630557497s" podCreationTimestamp="2025-10-04 08:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:18:49.629838889 +0000 UTC m=+157.384107703" watchObservedRunningTime="2025-10-04 08:18:49.630557497 +0000 UTC m=+157.384826311" Oct 04 08:18:49 crc kubenswrapper[4969]: I1004 08:18:49.666618 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:18:49 crc kubenswrapper[4969]: I1004 08:18:49.666676 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:18:50 crc kubenswrapper[4969]: I1004 08:18:50.617115 4969 generic.go:334] "Generic (PLEG): container finished" podID="2373497a-f79e-4f48-8792-4855301ff00d" containerID="7a2f1fe7792bcacc4efec54cfe9c2eb184d621852cc7e78c6a2a7050fd818e10" exitCode=0 Oct 04 08:18:50 crc kubenswrapper[4969]: I1004 08:18:50.617232 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2373497a-f79e-4f48-8792-4855301ff00d","Type":"ContainerDied","Data":"7a2f1fe7792bcacc4efec54cfe9c2eb184d621852cc7e78c6a2a7050fd818e10"} Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.050255 4969 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfqr6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.050344 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mfqr6" podUID="b4901cb1-78b0-4aa8-9e84-22d3c1f6669c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.050355 4969 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfqr6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.050497 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mfqr6" podUID="b4901cb1-78b0-4aa8-9e84-22d3c1f6669c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.080151 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.086295 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.236003 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.403350 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2373497a-f79e-4f48-8792-4855301ff00d-kube-api-access\") pod \"2373497a-f79e-4f48-8792-4855301ff00d\" (UID: \"2373497a-f79e-4f48-8792-4855301ff00d\") " Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.403515 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2373497a-f79e-4f48-8792-4855301ff00d-kubelet-dir\") pod \"2373497a-f79e-4f48-8792-4855301ff00d\" (UID: \"2373497a-f79e-4f48-8792-4855301ff00d\") " Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.403573 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2373497a-f79e-4f48-8792-4855301ff00d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2373497a-f79e-4f48-8792-4855301ff00d" (UID: "2373497a-f79e-4f48-8792-4855301ff00d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.403908 4969 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2373497a-f79e-4f48-8792-4855301ff00d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.408490 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2373497a-f79e-4f48-8792-4855301ff00d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2373497a-f79e-4f48-8792-4855301ff00d" (UID: "2373497a-f79e-4f48-8792-4855301ff00d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.505817 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2373497a-f79e-4f48-8792-4855301ff00d-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.647772 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2373497a-f79e-4f48-8792-4855301ff00d","Type":"ContainerDied","Data":"e52d9a631dbba75e65e43a9b55ec3990be505e884678baa5561dc3a0b2977cd5"} Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.647822 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e52d9a631dbba75e65e43a9b55ec3990be505e884678baa5561dc3a0b2977cd5" Oct 04 08:18:55 crc kubenswrapper[4969]: I1004 08:18:55.647843 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 08:18:56 crc kubenswrapper[4969]: I1004 08:18:56.620900 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:56 crc kubenswrapper[4969]: I1004 08:18:56.625853 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01650d78-a230-4e2d-80bb-817099bc34f2-metrics-certs\") pod \"network-metrics-daemon-kwwvg\" (UID: \"01650d78-a230-4e2d-80bb-817099bc34f2\") " pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:18:56 crc kubenswrapper[4969]: I1004 08:18:56.796919 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwwvg" Oct 04 08:19:03 crc kubenswrapper[4969]: I1004 08:19:03.619202 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:19:05 crc kubenswrapper[4969]: I1004 08:19:05.068205 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mfqr6" Oct 04 08:19:15 crc kubenswrapper[4969]: I1004 08:19:15.052784 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5mp4f" Oct 04 08:19:19 crc kubenswrapper[4969]: E1004 08:19:19.135927 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 08:19:19 crc kubenswrapper[4969]: E1004 08:19:19.136884 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-96b4z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zfxkq_openshift-marketplace(072c05bc-8cfa-40d0-ba2b-41af13739f3f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 08:19:19 crc kubenswrapper[4969]: E1004 08:19:19.138078 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-zfxkq" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" Oct 04 08:19:19 crc kubenswrapper[4969]: E1004 08:19:19.144525 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 08:19:19 crc kubenswrapper[4969]: E1004 08:19:19.145543 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rhgqn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rlrqv_openshift-marketplace(03b29be6-fac6-418e-97f1-a9cdad90fbf0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 08:19:19 crc kubenswrapper[4969]: E1004 08:19:19.146803 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rlrqv" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" Oct 04 08:19:19 crc kubenswrapper[4969]: I1004 08:19:19.503004 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kwwvg"] Oct 04 08:19:19 crc kubenswrapper[4969]: I1004 08:19:19.667106 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:19:19 crc kubenswrapper[4969]: I1004 08:19:19.667176 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:19:20 crc kubenswrapper[4969]: E1004 08:19:20.422247 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zfxkq" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" Oct 04 08:19:20 crc kubenswrapper[4969]: E1004 08:19:20.422302 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rlrqv" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" Oct 04 08:19:20 crc kubenswrapper[4969]: E1004 08:19:20.500856 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 08:19:20 crc kubenswrapper[4969]: E1004 08:19:20.501011 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bdczc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hvw9b_openshift-marketplace(74d8fc3b-4749-4d82-92d7-df0e63457fbe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 08:19:20 crc kubenswrapper[4969]: E1004 08:19:20.502189 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hvw9b" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" Oct 04 08:19:20 crc kubenswrapper[4969]: E1004 08:19:20.504217 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 08:19:20 crc kubenswrapper[4969]: E1004 08:19:20.504339 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b8kxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vpcfb_openshift-marketplace(55366ee0-2720-4e4a-83e3-c4fb9be1f6a9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 08:19:20 crc kubenswrapper[4969]: E1004 08:19:20.505637 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vpcfb" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" Oct 04 08:19:21 crc kubenswrapper[4969]: I1004 08:19:21.199022 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 08:19:23 crc kubenswrapper[4969]: E1004 08:19:23.480078 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 08:19:23 crc kubenswrapper[4969]: E1004 08:19:23.480981 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kpn9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-db8bm_openshift-marketplace(724eefc4-1c68-4593-8aaf-0144f80e3784): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 08:19:23 crc kubenswrapper[4969]: E1004 08:19:23.482341 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-db8bm" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" Oct 04 08:19:23 crc kubenswrapper[4969]: E1004 08:19:23.487196 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 08:19:23 crc kubenswrapper[4969]: E1004 08:19:23.487401 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vzzwv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-mxq79_openshift-marketplace(f0a988d3-5ee9-4833-97c8-05796123ff9a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 08:19:23 crc kubenswrapper[4969]: E1004 08:19:23.489225 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-mxq79" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" Oct 04 08:19:24 crc kubenswrapper[4969]: W1004 08:19:24.678960 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01650d78_a230_4e2d_80bb_817099bc34f2.slice/crio-7c2b7ca04c1a803fad3279c1b13a369085fb19c71753b37f4560acfd31d4b0f5 WatchSource:0}: Error finding container 7c2b7ca04c1a803fad3279c1b13a369085fb19c71753b37f4560acfd31d4b0f5: Status 404 returned error can't find the container with id 7c2b7ca04c1a803fad3279c1b13a369085fb19c71753b37f4560acfd31d4b0f5 Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.679539 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hvw9b" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.679585 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vpcfb" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.679782 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-db8bm" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.679868 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-mxq79" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.729768 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.729899 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lp6g7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-b9nrp_openshift-marketplace(46ed6532-e8c7-43c3-b80e-aeba1c5e2273): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.731212 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-b9nrp" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.771566 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.771789 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vcftw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-472gd_openshift-marketplace(b7a10e08-6a22-4d0a-af3a-b8bc8c234830): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.773206 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-472gd" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" Oct 04 08:19:24 crc kubenswrapper[4969]: I1004 08:19:24.811546 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" event={"ID":"01650d78-a230-4e2d-80bb-817099bc34f2","Type":"ContainerStarted","Data":"7c2b7ca04c1a803fad3279c1b13a369085fb19c71753b37f4560acfd31d4b0f5"} Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.813834 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-472gd" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" Oct 04 08:19:24 crc kubenswrapper[4969]: E1004 08:19:24.814277 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-b9nrp" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" Oct 04 08:19:25 crc kubenswrapper[4969]: I1004 08:19:25.817823 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" event={"ID":"01650d78-a230-4e2d-80bb-817099bc34f2","Type":"ContainerStarted","Data":"9b81cfbeea04a3bc83365db62f1de8550dd0182623cf51df8bd6ef65b3d6d24d"} Oct 04 08:19:25 crc kubenswrapper[4969]: I1004 08:19:25.818339 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kwwvg" event={"ID":"01650d78-a230-4e2d-80bb-817099bc34f2","Type":"ContainerStarted","Data":"94cc24d023ac33ea691f646a4020bd8c314b787e0f9e0d5121c68d541e77834d"} Oct 04 08:19:34 crc kubenswrapper[4969]: I1004 08:19:34.085791 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-kwwvg" podStartSLOduration=181.085722972 podStartE2EDuration="3m1.085722972s" podCreationTimestamp="2025-10-04 08:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:19:25.840528 +0000 UTC m=+193.594796864" watchObservedRunningTime="2025-10-04 08:19:34.085722972 +0000 UTC m=+201.839991816" Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.915182 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9nrp" event={"ID":"46ed6532-e8c7-43c3-b80e-aeba1c5e2273","Type":"ContainerStarted","Data":"e361fa8e23ae2122bb753bd71935ceef9f2582e0c453d4357914671cb4f8ef87"} Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.918153 4969 generic.go:334] "Generic (PLEG): container finished" podID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerID="e48e2341c4cb2495c33417a8b410b2bddaf7b030fe3cc5ce43f693bcdf8b5786" exitCode=0 Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.918245 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlrqv" event={"ID":"03b29be6-fac6-418e-97f1-a9cdad90fbf0","Type":"ContainerDied","Data":"e48e2341c4cb2495c33417a8b410b2bddaf7b030fe3cc5ce43f693bcdf8b5786"} Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.929732 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db8bm" event={"ID":"724eefc4-1c68-4593-8aaf-0144f80e3784","Type":"ContainerStarted","Data":"80ac2cdd245d389797a35a9a1af0e33adbacea0351d85b484d411525f820d588"} Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.932888 4969 generic.go:334] "Generic (PLEG): container finished" podID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerID="8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa" exitCode=0 Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.933173 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvw9b" event={"ID":"74d8fc3b-4749-4d82-92d7-df0e63457fbe","Type":"ContainerDied","Data":"8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa"} Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.943359 4969 generic.go:334] "Generic (PLEG): container finished" podID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerID="620acb639ea37cc8ddb372215f9591b088a11f8ac91abfe89fa97c4981278464" exitCode=0 Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.943483 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpcfb" event={"ID":"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9","Type":"ContainerDied","Data":"620acb639ea37cc8ddb372215f9591b088a11f8ac91abfe89fa97c4981278464"} Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.946223 4969 generic.go:334] "Generic (PLEG): container finished" podID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerID="a499cfdc4e192fff4f35e0cad0b0202a5fc6dbc2f92da95eb73de5f468c16a07" exitCode=0 Oct 04 08:19:41 crc kubenswrapper[4969]: I1004 08:19:41.946292 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfxkq" event={"ID":"072c05bc-8cfa-40d0-ba2b-41af13739f3f","Type":"ContainerDied","Data":"a499cfdc4e192fff4f35e0cad0b0202a5fc6dbc2f92da95eb73de5f468c16a07"} Oct 04 08:19:42 crc kubenswrapper[4969]: I1004 08:19:42.955347 4969 generic.go:334] "Generic (PLEG): container finished" podID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerID="6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6" exitCode=0 Oct 04 08:19:42 crc kubenswrapper[4969]: I1004 08:19:42.955445 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472gd" event={"ID":"b7a10e08-6a22-4d0a-af3a-b8bc8c234830","Type":"ContainerDied","Data":"6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6"} Oct 04 08:19:42 crc kubenswrapper[4969]: I1004 08:19:42.959825 4969 generic.go:334] "Generic (PLEG): container finished" podID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerID="f2b87ed3e0ff0a5d7c383cb4d6f354e8f49109c9ad588fc33985e95173c7c5d8" exitCode=0 Oct 04 08:19:42 crc kubenswrapper[4969]: I1004 08:19:42.959923 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxq79" event={"ID":"f0a988d3-5ee9-4833-97c8-05796123ff9a","Type":"ContainerDied","Data":"f2b87ed3e0ff0a5d7c383cb4d6f354e8f49109c9ad588fc33985e95173c7c5d8"} Oct 04 08:19:42 crc kubenswrapper[4969]: I1004 08:19:42.963240 4969 generic.go:334] "Generic (PLEG): container finished" podID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerID="e361fa8e23ae2122bb753bd71935ceef9f2582e0c453d4357914671cb4f8ef87" exitCode=0 Oct 04 08:19:42 crc kubenswrapper[4969]: I1004 08:19:42.963280 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9nrp" event={"ID":"46ed6532-e8c7-43c3-b80e-aeba1c5e2273","Type":"ContainerDied","Data":"e361fa8e23ae2122bb753bd71935ceef9f2582e0c453d4357914671cb4f8ef87"} Oct 04 08:19:42 crc kubenswrapper[4969]: I1004 08:19:42.966123 4969 generic.go:334] "Generic (PLEG): container finished" podID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerID="80ac2cdd245d389797a35a9a1af0e33adbacea0351d85b484d411525f820d588" exitCode=0 Oct 04 08:19:42 crc kubenswrapper[4969]: I1004 08:19:42.966156 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db8bm" event={"ID":"724eefc4-1c68-4593-8aaf-0144f80e3784","Type":"ContainerDied","Data":"80ac2cdd245d389797a35a9a1af0e33adbacea0351d85b484d411525f820d588"} Oct 04 08:19:43 crc kubenswrapper[4969]: I1004 08:19:43.981477 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxq79" event={"ID":"f0a988d3-5ee9-4833-97c8-05796123ff9a","Type":"ContainerStarted","Data":"d22ac309ab7c2844bf78fb7bf3c6c082207fb5f41cbfd59c401360dba1bf8fc3"} Oct 04 08:19:43 crc kubenswrapper[4969]: I1004 08:19:43.987337 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvw9b" event={"ID":"74d8fc3b-4749-4d82-92d7-df0e63457fbe","Type":"ContainerStarted","Data":"5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01"} Oct 04 08:19:43 crc kubenswrapper[4969]: I1004 08:19:43.991957 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpcfb" event={"ID":"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9","Type":"ContainerStarted","Data":"fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7"} Oct 04 08:19:43 crc kubenswrapper[4969]: I1004 08:19:43.994861 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfxkq" event={"ID":"072c05bc-8cfa-40d0-ba2b-41af13739f3f","Type":"ContainerStarted","Data":"ab9161e9a12a4754a5c751ccf7d5657799d1a9887ea48318811696aebd169510"} Oct 04 08:19:43 crc kubenswrapper[4969]: I1004 08:19:43.997002 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlrqv" event={"ID":"03b29be6-fac6-418e-97f1-a9cdad90fbf0","Type":"ContainerStarted","Data":"f31f51c0d54a4c6486ee1f841f9a4c6cd7da38de8bb7c5c201ab3a5726b21a88"} Oct 04 08:19:43 crc kubenswrapper[4969]: I1004 08:19:43.998559 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db8bm" event={"ID":"724eefc4-1c68-4593-8aaf-0144f80e3784","Type":"ContainerStarted","Data":"0babd9b9326efd8acf17538041edde83c169585697d588c7f150d75ecbaf4a54"} Oct 04 08:19:44 crc kubenswrapper[4969]: I1004 08:19:44.001481 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472gd" event={"ID":"b7a10e08-6a22-4d0a-af3a-b8bc8c234830","Type":"ContainerStarted","Data":"196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151"} Oct 04 08:19:44 crc kubenswrapper[4969]: I1004 08:19:44.006814 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mxq79" podStartSLOduration=2.846990342 podStartE2EDuration="1m3.006795643s" podCreationTimestamp="2025-10-04 08:18:41 +0000 UTC" firstStartedPulling="2025-10-04 08:18:43.404039101 +0000 UTC m=+151.158307915" lastFinishedPulling="2025-10-04 08:19:43.563844402 +0000 UTC m=+211.318113216" observedRunningTime="2025-10-04 08:19:44.003283661 +0000 UTC m=+211.757552485" watchObservedRunningTime="2025-10-04 08:19:44.006795643 +0000 UTC m=+211.761064477" Oct 04 08:19:44 crc kubenswrapper[4969]: I1004 08:19:44.024632 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zfxkq" podStartSLOduration=3.148400626 podStartE2EDuration="1m4.024611901s" podCreationTimestamp="2025-10-04 08:18:40 +0000 UTC" firstStartedPulling="2025-10-04 08:18:42.388695015 +0000 UTC m=+150.142963829" lastFinishedPulling="2025-10-04 08:19:43.26490629 +0000 UTC m=+211.019175104" observedRunningTime="2025-10-04 08:19:44.022745142 +0000 UTC m=+211.777013956" watchObservedRunningTime="2025-10-04 08:19:44.024611901 +0000 UTC m=+211.778880725" Oct 04 08:19:44 crc kubenswrapper[4969]: I1004 08:19:44.084755 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vpcfb" podStartSLOduration=3.938561484 podStartE2EDuration="1m7.084737528s" podCreationTimestamp="2025-10-04 08:18:37 +0000 UTC" firstStartedPulling="2025-10-04 08:18:39.915090862 +0000 UTC m=+147.669359676" lastFinishedPulling="2025-10-04 08:19:43.061266906 +0000 UTC m=+210.815535720" observedRunningTime="2025-10-04 08:19:44.057821742 +0000 UTC m=+211.812090556" watchObservedRunningTime="2025-10-04 08:19:44.084737528 +0000 UTC m=+211.839006342" Oct 04 08:19:44 crc kubenswrapper[4969]: I1004 08:19:44.085981 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rlrqv" podStartSLOduration=4.271149167 podStartE2EDuration="1m5.085969781s" podCreationTimestamp="2025-10-04 08:18:39 +0000 UTC" firstStartedPulling="2025-10-04 08:18:42.33938584 +0000 UTC m=+150.093654644" lastFinishedPulling="2025-10-04 08:19:43.154206444 +0000 UTC m=+210.908475258" observedRunningTime="2025-10-04 08:19:44.073459932 +0000 UTC m=+211.827728746" watchObservedRunningTime="2025-10-04 08:19:44.085969781 +0000 UTC m=+211.840238595" Oct 04 08:19:44 crc kubenswrapper[4969]: I1004 08:19:44.091777 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hvw9b" podStartSLOduration=4.013081884 podStartE2EDuration="1m6.091757232s" podCreationTimestamp="2025-10-04 08:18:38 +0000 UTC" firstStartedPulling="2025-10-04 08:18:41.252565701 +0000 UTC m=+149.006834515" lastFinishedPulling="2025-10-04 08:19:43.331241049 +0000 UTC m=+211.085509863" observedRunningTime="2025-10-04 08:19:44.088776704 +0000 UTC m=+211.843045518" watchObservedRunningTime="2025-10-04 08:19:44.091757232 +0000 UTC m=+211.846026046" Oct 04 08:19:44 crc kubenswrapper[4969]: I1004 08:19:44.114229 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-472gd" podStartSLOduration=3.830536298 podStartE2EDuration="1m6.114211671s" podCreationTimestamp="2025-10-04 08:18:38 +0000 UTC" firstStartedPulling="2025-10-04 08:18:41.207040201 +0000 UTC m=+148.961309015" lastFinishedPulling="2025-10-04 08:19:43.490715564 +0000 UTC m=+211.244984388" observedRunningTime="2025-10-04 08:19:44.113722519 +0000 UTC m=+211.867991333" watchObservedRunningTime="2025-10-04 08:19:44.114211671 +0000 UTC m=+211.868480485" Oct 04 08:19:44 crc kubenswrapper[4969]: I1004 08:19:44.152632 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-db8bm" podStartSLOduration=3.167906368 podStartE2EDuration="1m3.15261655s" podCreationTimestamp="2025-10-04 08:18:41 +0000 UTC" firstStartedPulling="2025-10-04 08:18:43.390494591 +0000 UTC m=+151.144763405" lastFinishedPulling="2025-10-04 08:19:43.375204743 +0000 UTC m=+211.129473587" observedRunningTime="2025-10-04 08:19:44.145009439 +0000 UTC m=+211.899278263" watchObservedRunningTime="2025-10-04 08:19:44.15261655 +0000 UTC m=+211.906885354" Oct 04 08:19:45 crc kubenswrapper[4969]: I1004 08:19:45.007386 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9nrp" event={"ID":"46ed6532-e8c7-43c3-b80e-aeba1c5e2273","Type":"ContainerStarted","Data":"326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d"} Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.228176 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.228750 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.481524 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.481841 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.520923 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.528906 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.538816 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b9nrp" podStartSLOduration=7.782441277 podStartE2EDuration="1m10.538799345s" podCreationTimestamp="2025-10-04 08:18:38 +0000 UTC" firstStartedPulling="2025-10-04 08:18:41.27082009 +0000 UTC m=+149.025088904" lastFinishedPulling="2025-10-04 08:19:44.027178148 +0000 UTC m=+211.781446972" observedRunningTime="2025-10-04 08:19:45.032318599 +0000 UTC m=+212.786587413" watchObservedRunningTime="2025-10-04 08:19:48.538799345 +0000 UTC m=+216.293068159" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.610794 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.610862 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.655170 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.872091 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.872170 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:19:48 crc kubenswrapper[4969]: I1004 08:19:48.928087 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.066975 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.072664 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.078217 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.086711 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.666360 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.666401 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.666456 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.666870 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.666947 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452" gracePeriod=600 Oct 04 08:19:49 crc kubenswrapper[4969]: I1004 08:19:49.887922 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hvw9b"] Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.032580 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452" exitCode=0 Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.032673 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452"} Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.033058 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"d265f04fbe19dd6f5ad0bc22c79d5d325f5680917d64f171975ad5c10938d825"} Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.188587 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.188640 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.230796 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.705164 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.705215 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.753928 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:19:50 crc kubenswrapper[4969]: I1004 08:19:50.883669 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-472gd"] Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.037642 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hvw9b" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerName="registry-server" containerID="cri-o://5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01" gracePeriod=2 Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.039298 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-472gd" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerName="registry-server" containerID="cri-o://196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151" gracePeriod=2 Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.105655 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.109411 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.392522 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.418471 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.418806 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.435157 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.437830 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-catalog-content\") pod \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.437924 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcftw\" (UniqueName: \"kubernetes.io/projected/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-kube-api-access-vcftw\") pod \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.437951 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-utilities\") pod \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\" (UID: \"b7a10e08-6a22-4d0a-af3a-b8bc8c234830\") " Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.443440 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-utilities" (OuterVolumeSpecName: "utilities") pod "b7a10e08-6a22-4d0a-af3a-b8bc8c234830" (UID: "b7a10e08-6a22-4d0a-af3a-b8bc8c234830"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.446032 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-kube-api-access-vcftw" (OuterVolumeSpecName: "kube-api-access-vcftw") pod "b7a10e08-6a22-4d0a-af3a-b8bc8c234830" (UID: "b7a10e08-6a22-4d0a-af3a-b8bc8c234830"). InnerVolumeSpecName "kube-api-access-vcftw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.463213 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.528937 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7a10e08-6a22-4d0a-af3a-b8bc8c234830" (UID: "b7a10e08-6a22-4d0a-af3a-b8bc8c234830"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.538467 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdczc\" (UniqueName: \"kubernetes.io/projected/74d8fc3b-4749-4d82-92d7-df0e63457fbe-kube-api-access-bdczc\") pod \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.538520 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-catalog-content\") pod \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.538624 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-utilities\") pod \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\" (UID: \"74d8fc3b-4749-4d82-92d7-df0e63457fbe\") " Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.538828 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.538841 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcftw\" (UniqueName: \"kubernetes.io/projected/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-kube-api-access-vcftw\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.538851 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a10e08-6a22-4d0a-af3a-b8bc8c234830-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.539337 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-utilities" (OuterVolumeSpecName: "utilities") pod "74d8fc3b-4749-4d82-92d7-df0e63457fbe" (UID: "74d8fc3b-4749-4d82-92d7-df0e63457fbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.542030 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74d8fc3b-4749-4d82-92d7-df0e63457fbe-kube-api-access-bdczc" (OuterVolumeSpecName: "kube-api-access-bdczc") pod "74d8fc3b-4749-4d82-92d7-df0e63457fbe" (UID: "74d8fc3b-4749-4d82-92d7-df0e63457fbe"). InnerVolumeSpecName "kube-api-access-bdczc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.594986 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74d8fc3b-4749-4d82-92d7-df0e63457fbe" (UID: "74d8fc3b-4749-4d82-92d7-df0e63457fbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.639641 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.639675 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdczc\" (UniqueName: \"kubernetes.io/projected/74d8fc3b-4749-4d82-92d7-df0e63457fbe-kube-api-access-bdczc\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.639687 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d8fc3b-4749-4d82-92d7-df0e63457fbe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.850284 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.850618 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:19:51 crc kubenswrapper[4969]: I1004 08:19:51.891386 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.046489 4969 generic.go:334] "Generic (PLEG): container finished" podID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerID="196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151" exitCode=0 Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.046551 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472gd" event={"ID":"b7a10e08-6a22-4d0a-af3a-b8bc8c234830","Type":"ContainerDied","Data":"196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151"} Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.046579 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472gd" event={"ID":"b7a10e08-6a22-4d0a-af3a-b8bc8c234830","Type":"ContainerDied","Data":"bf669b959e3f52c02f28f4ea079b2f01b7c387e2d78148182a74cbc70e60a7f5"} Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.046597 4969 scope.go:117] "RemoveContainer" containerID="196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.046593 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-472gd" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.048938 4969 generic.go:334] "Generic (PLEG): container finished" podID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerID="5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01" exitCode=0 Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.049010 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvw9b" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.049080 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvw9b" event={"ID":"74d8fc3b-4749-4d82-92d7-df0e63457fbe","Type":"ContainerDied","Data":"5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01"} Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.049110 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvw9b" event={"ID":"74d8fc3b-4749-4d82-92d7-df0e63457fbe","Type":"ContainerDied","Data":"a16e838189f8c416132621aeaad5e9f8b2c7b70df6f369bb6ce201935d53675f"} Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.064140 4969 scope.go:117] "RemoveContainer" containerID="6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.079874 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hvw9b"] Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.087063 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hvw9b"] Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.087623 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.087839 4969 scope.go:117] "RemoveContainer" containerID="532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.091241 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.092249 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-472gd"] Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.095382 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-472gd"] Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.101873 4969 scope.go:117] "RemoveContainer" containerID="196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151" Oct 04 08:19:52 crc kubenswrapper[4969]: E1004 08:19:52.104654 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151\": container with ID starting with 196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151 not found: ID does not exist" containerID="196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.104691 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151"} err="failed to get container status \"196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151\": rpc error: code = NotFound desc = could not find container \"196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151\": container with ID starting with 196f403ad8e27c88fb907a70c6a9d1b1adb0267d9508c750bee09adafbfa4151 not found: ID does not exist" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.104717 4969 scope.go:117] "RemoveContainer" containerID="6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6" Oct 04 08:19:52 crc kubenswrapper[4969]: E1004 08:19:52.105190 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6\": container with ID starting with 6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6 not found: ID does not exist" containerID="6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.105233 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6"} err="failed to get container status \"6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6\": rpc error: code = NotFound desc = could not find container \"6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6\": container with ID starting with 6135bdbe020b859c552fc77550391ae5054fd0c84174b6b95687f9a62fc267f6 not found: ID does not exist" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.105290 4969 scope.go:117] "RemoveContainer" containerID="532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8" Oct 04 08:19:52 crc kubenswrapper[4969]: E1004 08:19:52.105748 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8\": container with ID starting with 532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8 not found: ID does not exist" containerID="532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.105777 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8"} err="failed to get container status \"532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8\": rpc error: code = NotFound desc = could not find container \"532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8\": container with ID starting with 532d12d7f320f64e969fdd169ac83aa3f7634b39a92e749d695249b28197c0b8 not found: ID does not exist" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.105796 4969 scope.go:117] "RemoveContainer" containerID="5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.122877 4969 scope.go:117] "RemoveContainer" containerID="8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.137778 4969 scope.go:117] "RemoveContainer" containerID="5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.150103 4969 scope.go:117] "RemoveContainer" containerID="5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01" Oct 04 08:19:52 crc kubenswrapper[4969]: E1004 08:19:52.150475 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01\": container with ID starting with 5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01 not found: ID does not exist" containerID="5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.150505 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01"} err="failed to get container status \"5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01\": rpc error: code = NotFound desc = could not find container \"5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01\": container with ID starting with 5c81c639ca27dde61ce3b20612cf29234f4b4ea6300506d995eea33b4c1d7e01 not found: ID does not exist" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.150525 4969 scope.go:117] "RemoveContainer" containerID="8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa" Oct 04 08:19:52 crc kubenswrapper[4969]: E1004 08:19:52.150938 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa\": container with ID starting with 8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa not found: ID does not exist" containerID="8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.150975 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa"} err="failed to get container status \"8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa\": rpc error: code = NotFound desc = could not find container \"8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa\": container with ID starting with 8fd0308d0ef0ca828a11717b0b7d3ea916792833a816c9f2d915a305489a81aa not found: ID does not exist" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.151002 4969 scope.go:117] "RemoveContainer" containerID="5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d" Oct 04 08:19:52 crc kubenswrapper[4969]: E1004 08:19:52.151315 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d\": container with ID starting with 5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d not found: ID does not exist" containerID="5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d" Oct 04 08:19:52 crc kubenswrapper[4969]: I1004 08:19:52.151342 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d"} err="failed to get container status \"5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d\": rpc error: code = NotFound desc = could not find container \"5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d\": container with ID starting with 5be9b65fbba051eb3b701d0b29ef153e05217b495fde0ea75da07305eec6284d not found: ID does not exist" Oct 04 08:19:53 crc kubenswrapper[4969]: I1004 08:19:53.062788 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" path="/var/lib/kubelet/pods/74d8fc3b-4749-4d82-92d7-df0e63457fbe/volumes" Oct 04 08:19:53 crc kubenswrapper[4969]: I1004 08:19:53.063372 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" path="/var/lib/kubelet/pods/b7a10e08-6a22-4d0a-af3a-b8bc8c234830/volumes" Oct 04 08:19:53 crc kubenswrapper[4969]: I1004 08:19:53.286316 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfxkq"] Oct 04 08:19:53 crc kubenswrapper[4969]: I1004 08:19:53.286582 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zfxkq" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerName="registry-server" containerID="cri-o://ab9161e9a12a4754a5c751ccf7d5657799d1a9887ea48318811696aebd169510" gracePeriod=2 Oct 04 08:19:54 crc kubenswrapper[4969]: I1004 08:19:54.062768 4969 generic.go:334] "Generic (PLEG): container finished" podID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerID="ab9161e9a12a4754a5c751ccf7d5657799d1a9887ea48318811696aebd169510" exitCode=0 Oct 04 08:19:54 crc kubenswrapper[4969]: I1004 08:19:54.062840 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfxkq" event={"ID":"072c05bc-8cfa-40d0-ba2b-41af13739f3f","Type":"ContainerDied","Data":"ab9161e9a12a4754a5c751ccf7d5657799d1a9887ea48318811696aebd169510"} Oct 04 08:19:54 crc kubenswrapper[4969]: I1004 08:19:54.775059 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:19:54 crc kubenswrapper[4969]: I1004 08:19:54.903989 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-utilities\") pod \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " Oct 04 08:19:54 crc kubenswrapper[4969]: I1004 08:19:54.904404 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96b4z\" (UniqueName: \"kubernetes.io/projected/072c05bc-8cfa-40d0-ba2b-41af13739f3f-kube-api-access-96b4z\") pod \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " Oct 04 08:19:54 crc kubenswrapper[4969]: I1004 08:19:54.904463 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-catalog-content\") pod \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\" (UID: \"072c05bc-8cfa-40d0-ba2b-41af13739f3f\") " Oct 04 08:19:54 crc kubenswrapper[4969]: I1004 08:19:54.906060 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-utilities" (OuterVolumeSpecName: "utilities") pod "072c05bc-8cfa-40d0-ba2b-41af13739f3f" (UID: "072c05bc-8cfa-40d0-ba2b-41af13739f3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:19:54 crc kubenswrapper[4969]: I1004 08:19:54.916104 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072c05bc-8cfa-40d0-ba2b-41af13739f3f-kube-api-access-96b4z" (OuterVolumeSpecName: "kube-api-access-96b4z") pod "072c05bc-8cfa-40d0-ba2b-41af13739f3f" (UID: "072c05bc-8cfa-40d0-ba2b-41af13739f3f"). InnerVolumeSpecName "kube-api-access-96b4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:19:54 crc kubenswrapper[4969]: I1004 08:19:54.916935 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "072c05bc-8cfa-40d0-ba2b-41af13739f3f" (UID: "072c05bc-8cfa-40d0-ba2b-41af13739f3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.005823 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.005863 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072c05bc-8cfa-40d0-ba2b-41af13739f3f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.005877 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96b4z\" (UniqueName: \"kubernetes.io/projected/072c05bc-8cfa-40d0-ba2b-41af13739f3f-kube-api-access-96b4z\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.068569 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfxkq" event={"ID":"072c05bc-8cfa-40d0-ba2b-41af13739f3f","Type":"ContainerDied","Data":"0c7e57e11ba180d9da2f049379d8a12fa3b01038fe90e8e8cc3e40bde15eface"} Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.068643 4969 scope.go:117] "RemoveContainer" containerID="ab9161e9a12a4754a5c751ccf7d5657799d1a9887ea48318811696aebd169510" Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.068592 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfxkq" Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.103570 4969 scope.go:117] "RemoveContainer" containerID="a499cfdc4e192fff4f35e0cad0b0202a5fc6dbc2f92da95eb73de5f468c16a07" Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.119025 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfxkq"] Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.119918 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfxkq"] Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.141082 4969 scope.go:117] "RemoveContainer" containerID="83aaabc54fac748acc33f873266b06f0555915c22b5b6f1ed76cee2ee3e15f64" Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.684886 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-db8bm"] Oct 04 08:19:55 crc kubenswrapper[4969]: I1004 08:19:55.685386 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-db8bm" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerName="registry-server" containerID="cri-o://0babd9b9326efd8acf17538041edde83c169585697d588c7f150d75ecbaf4a54" gracePeriod=2 Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.062627 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" path="/var/lib/kubelet/pods/072c05bc-8cfa-40d0-ba2b-41af13739f3f/volumes" Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.081912 4969 generic.go:334] "Generic (PLEG): container finished" podID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerID="0babd9b9326efd8acf17538041edde83c169585697d588c7f150d75ecbaf4a54" exitCode=0 Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.081949 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db8bm" event={"ID":"724eefc4-1c68-4593-8aaf-0144f80e3784","Type":"ContainerDied","Data":"0babd9b9326efd8acf17538041edde83c169585697d588c7f150d75ecbaf4a54"} Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.473384 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.668082 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-catalog-content\") pod \"724eefc4-1c68-4593-8aaf-0144f80e3784\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.668137 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-utilities\") pod \"724eefc4-1c68-4593-8aaf-0144f80e3784\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.668166 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpn9q\" (UniqueName: \"kubernetes.io/projected/724eefc4-1c68-4593-8aaf-0144f80e3784-kube-api-access-kpn9q\") pod \"724eefc4-1c68-4593-8aaf-0144f80e3784\" (UID: \"724eefc4-1c68-4593-8aaf-0144f80e3784\") " Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.670063 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-utilities" (OuterVolumeSpecName: "utilities") pod "724eefc4-1c68-4593-8aaf-0144f80e3784" (UID: "724eefc4-1c68-4593-8aaf-0144f80e3784"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.675389 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/724eefc4-1c68-4593-8aaf-0144f80e3784-kube-api-access-kpn9q" (OuterVolumeSpecName: "kube-api-access-kpn9q") pod "724eefc4-1c68-4593-8aaf-0144f80e3784" (UID: "724eefc4-1c68-4593-8aaf-0144f80e3784"). InnerVolumeSpecName "kube-api-access-kpn9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.765176 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "724eefc4-1c68-4593-8aaf-0144f80e3784" (UID: "724eefc4-1c68-4593-8aaf-0144f80e3784"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.769692 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.769725 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724eefc4-1c68-4593-8aaf-0144f80e3784-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:57 crc kubenswrapper[4969]: I1004 08:19:57.769737 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpn9q\" (UniqueName: \"kubernetes.io/projected/724eefc4-1c68-4593-8aaf-0144f80e3784-kube-api-access-kpn9q\") on node \"crc\" DevicePath \"\"" Oct 04 08:19:58 crc kubenswrapper[4969]: I1004 08:19:58.088767 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db8bm" event={"ID":"724eefc4-1c68-4593-8aaf-0144f80e3784","Type":"ContainerDied","Data":"2ea5fdb2c2afbb1970f7d6213bff7b53734f071b6b627015cd3c8c79a99018a3"} Oct 04 08:19:58 crc kubenswrapper[4969]: I1004 08:19:58.088825 4969 scope.go:117] "RemoveContainer" containerID="0babd9b9326efd8acf17538041edde83c169585697d588c7f150d75ecbaf4a54" Oct 04 08:19:58 crc kubenswrapper[4969]: I1004 08:19:58.088828 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db8bm" Oct 04 08:19:58 crc kubenswrapper[4969]: I1004 08:19:58.104803 4969 scope.go:117] "RemoveContainer" containerID="80ac2cdd245d389797a35a9a1af0e33adbacea0351d85b484d411525f820d588" Oct 04 08:19:58 crc kubenswrapper[4969]: I1004 08:19:58.113301 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-db8bm"] Oct 04 08:19:58 crc kubenswrapper[4969]: I1004 08:19:58.120621 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-db8bm"] Oct 04 08:19:58 crc kubenswrapper[4969]: I1004 08:19:58.128182 4969 scope.go:117] "RemoveContainer" containerID="3f225ef56d508ec4094f2830bc4ae2b8c9a88d142c7da2ea693f46e09fd3aed3" Oct 04 08:19:59 crc kubenswrapper[4969]: I1004 08:19:59.062057 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" path="/var/lib/kubelet/pods/724eefc4-1c68-4593-8aaf-0144f80e3784/volumes" Oct 04 08:20:04 crc kubenswrapper[4969]: I1004 08:20:04.719510 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8xngh"] Oct 04 08:20:29 crc kubenswrapper[4969]: I1004 08:20:29.762124 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" podUID="7f61cf73-65f6-4255-8c99-4dfd6d48d80c" containerName="oauth-openshift" containerID="cri-o://ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac" gracePeriod=15 Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.135499 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186354 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-l25zj"] Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186608 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f61cf73-65f6-4255-8c99-4dfd6d48d80c" containerName="oauth-openshift" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186623 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f61cf73-65f6-4255-8c99-4dfd6d48d80c" containerName="oauth-openshift" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186635 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerName="extract-content" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186644 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerName="extract-content" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186658 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186666 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186674 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186681 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186693 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2373497a-f79e-4f48-8792-4855301ff00d" containerName="pruner" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186702 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="2373497a-f79e-4f48-8792-4855301ff00d" containerName="pruner" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186714 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186722 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186736 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerName="extract-content" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186744 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerName="extract-content" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186756 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerName="extract-content" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186764 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerName="extract-content" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186776 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerName="extract-utilities" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186784 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerName="extract-utilities" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186795 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerName="extract-content" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186802 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerName="extract-content" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186813 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186821 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186836 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerName="extract-utilities" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186843 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerName="extract-utilities" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186854 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerName="extract-utilities" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186862 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerName="extract-utilities" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.186873 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerName="extract-utilities" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186880 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerName="extract-utilities" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.186993 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7a10e08-6a22-4d0a-af3a-b8bc8c234830" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.187005 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="724eefc4-1c68-4593-8aaf-0144f80e3784" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.187019 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="74d8fc3b-4749-4d82-92d7-df0e63457fbe" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.187030 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="2373497a-f79e-4f48-8792-4855301ff00d" containerName="pruner" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.187037 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="072c05bc-8cfa-40d0-ba2b-41af13739f3f" containerName="registry-server" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.187048 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f61cf73-65f6-4255-8c99-4dfd6d48d80c" containerName="oauth-openshift" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.187451 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.195517 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-l25zj"] Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253196 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-session\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253266 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-error\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253306 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-cliconfig\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253328 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-dir\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253357 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253835 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-login\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253884 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-router-certs\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253913 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-service-ca\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253935 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-policies\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253958 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-idp-0-file-data\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.253988 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-ocp-branding-template\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254013 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-serving-cert\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254044 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmrx7\" (UniqueName: \"kubernetes.io/projected/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-kube-api-access-lmrx7\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254074 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-trusted-ca-bundle\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254131 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-provider-selection\") pod \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\" (UID: \"7f61cf73-65f6-4255-8c99-4dfd6d48d80c\") " Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254259 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254302 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254344 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-session\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254380 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-router-certs\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254454 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-service-ca\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254494 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254520 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254525 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254560 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254596 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae754295-a052-4f94-bac3-4c83b0994ca1-audit-dir\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254631 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254661 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254698 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-audit-policies\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254733 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x9w4\" (UniqueName: \"kubernetes.io/projected/ae754295-a052-4f94-bac3-4c83b0994ca1-kube-api-access-8x9w4\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254766 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-template-login\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254804 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-template-error\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254874 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254898 4969 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254921 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.254925 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.255160 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.259221 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.259593 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.259764 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-kube-api-access-lmrx7" (OuterVolumeSpecName: "kube-api-access-lmrx7") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "kube-api-access-lmrx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.259933 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.260375 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.260766 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.261155 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.264988 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.265213 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7f61cf73-65f6-4255-8c99-4dfd6d48d80c" (UID: "7f61cf73-65f6-4255-8c99-4dfd6d48d80c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.284566 4969 generic.go:334] "Generic (PLEG): container finished" podID="7f61cf73-65f6-4255-8c99-4dfd6d48d80c" containerID="ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac" exitCode=0 Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.284600 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" event={"ID":"7f61cf73-65f6-4255-8c99-4dfd6d48d80c","Type":"ContainerDied","Data":"ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac"} Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.284677 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" event={"ID":"7f61cf73-65f6-4255-8c99-4dfd6d48d80c","Type":"ContainerDied","Data":"5dc12655eefaa97786c515da5530a463eae91587ef1f0c66f44b508abc2e00ec"} Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.284698 4969 scope.go:117] "RemoveContainer" containerID="ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.284987 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8xngh" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.312203 4969 scope.go:117] "RemoveContainer" containerID="ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac" Oct 04 08:20:30 crc kubenswrapper[4969]: E1004 08:20:30.315038 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac\": container with ID starting with ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac not found: ID does not exist" containerID="ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.315077 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac"} err="failed to get container status \"ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac\": rpc error: code = NotFound desc = could not find container \"ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac\": container with ID starting with ce1a19f4d20d417b56e5ecba0c1b4855a6602863da5b572482a84587fb86a9ac not found: ID does not exist" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.318945 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8xngh"] Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.322599 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8xngh"] Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.355928 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.356250 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-audit-policies\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.356463 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x9w4\" (UniqueName: \"kubernetes.io/projected/ae754295-a052-4f94-bac3-4c83b0994ca1-kube-api-access-8x9w4\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.356672 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-template-login\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.356872 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-template-error\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.357048 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.357246 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.357479 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-session\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.357656 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-router-certs\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.357822 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-service-ca\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.357999 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.358127 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.358240 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae754295-a052-4f94-bac3-4c83b0994ca1-audit-dir\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.358341 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.358338 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae754295-a052-4f94-bac3-4c83b0994ca1-audit-dir\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.357497 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-audit-policies\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.358718 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.358753 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.359758 4969 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.359922 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.360030 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.359860 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-template-error\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.359885 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-service-ca\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.359721 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.360116 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.360211 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmrx7\" (UniqueName: \"kubernetes.io/projected/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-kube-api-access-lmrx7\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.360235 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.360260 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.360277 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.359555 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.360294 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.360312 4969 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f61cf73-65f6-4255-8c99-4dfd6d48d80c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.362155 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-template-login\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.363098 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.364470 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.365375 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-router-certs\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.365765 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-system-session\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.366540 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae754295-a052-4f94-bac3-4c83b0994ca1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.381490 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x9w4\" (UniqueName: \"kubernetes.io/projected/ae754295-a052-4f94-bac3-4c83b0994ca1-kube-api-access-8x9w4\") pod \"oauth-openshift-64f9fb64bf-l25zj\" (UID: \"ae754295-a052-4f94-bac3-4c83b0994ca1\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.518525 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:30 crc kubenswrapper[4969]: I1004 08:20:30.799962 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-l25zj"] Oct 04 08:20:31 crc kubenswrapper[4969]: I1004 08:20:31.064471 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f61cf73-65f6-4255-8c99-4dfd6d48d80c" path="/var/lib/kubelet/pods/7f61cf73-65f6-4255-8c99-4dfd6d48d80c/volumes" Oct 04 08:20:31 crc kubenswrapper[4969]: I1004 08:20:31.297539 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" event={"ID":"ae754295-a052-4f94-bac3-4c83b0994ca1","Type":"ContainerStarted","Data":"a793f57c92aa1f6e0c37ada9454850f88c36cad4ebc63d9dcafe31c01849ac38"} Oct 04 08:20:31 crc kubenswrapper[4969]: I1004 08:20:31.297823 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" event={"ID":"ae754295-a052-4f94-bac3-4c83b0994ca1","Type":"ContainerStarted","Data":"78bd9d28a95f155d88d2fe4bd1c32a36682ffce7372326e5ccc277c258087b57"} Oct 04 08:20:31 crc kubenswrapper[4969]: I1004 08:20:31.297843 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:31 crc kubenswrapper[4969]: I1004 08:20:31.848766 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" Oct 04 08:20:31 crc kubenswrapper[4969]: I1004 08:20:31.871478 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-64f9fb64bf-l25zj" podStartSLOduration=27.871458493 podStartE2EDuration="27.871458493s" podCreationTimestamp="2025-10-04 08:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:20:31.321850266 +0000 UTC m=+259.076119140" watchObservedRunningTime="2025-10-04 08:20:31.871458493 +0000 UTC m=+259.625727307" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.056439 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9nrp"] Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.057830 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b9nrp" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerName="registry-server" containerID="cri-o://326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d" gracePeriod=30 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.064360 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vpcfb"] Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.064700 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vpcfb" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerName="registry-server" containerID="cri-o://fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7" gracePeriod=30 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.078522 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7tnxs"] Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.078759 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" podUID="d0a906da-f3d3-40cf-a714-2159253bb043" containerName="marketplace-operator" containerID="cri-o://f50e9dce1265bd029868849547362f4dd15bfd5fdccb1f64cb08a3a8c76d7448" gracePeriod=30 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.083949 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlrqv"] Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.084207 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rlrqv" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerName="registry-server" containerID="cri-o://f31f51c0d54a4c6486ee1f841f9a4c6cd7da38de8bb7c5c201ab3a5726b21a88" gracePeriod=30 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.092725 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5qth5"] Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.093721 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.100376 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxq79"] Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.102026 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mxq79" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerName="registry-server" containerID="cri-o://d22ac309ab7c2844bf78fb7bf3c6c082207fb5f41cbfd59c401360dba1bf8fc3" gracePeriod=30 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.107176 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9wbt\" (UniqueName: \"kubernetes.io/projected/0cf38a30-f714-4a15-8be5-30118e8984c7-kube-api-access-t9wbt\") pod \"marketplace-operator-79b997595-5qth5\" (UID: \"0cf38a30-f714-4a15-8be5-30118e8984c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.107452 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0cf38a30-f714-4a15-8be5-30118e8984c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5qth5\" (UID: \"0cf38a30-f714-4a15-8be5-30118e8984c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.107492 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cf38a30-f714-4a15-8be5-30118e8984c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5qth5\" (UID: \"0cf38a30-f714-4a15-8be5-30118e8984c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.111991 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5qth5"] Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.208842 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cf38a30-f714-4a15-8be5-30118e8984c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5qth5\" (UID: \"0cf38a30-f714-4a15-8be5-30118e8984c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.209035 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9wbt\" (UniqueName: \"kubernetes.io/projected/0cf38a30-f714-4a15-8be5-30118e8984c7-kube-api-access-t9wbt\") pod \"marketplace-operator-79b997595-5qth5\" (UID: \"0cf38a30-f714-4a15-8be5-30118e8984c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.209094 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0cf38a30-f714-4a15-8be5-30118e8984c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5qth5\" (UID: \"0cf38a30-f714-4a15-8be5-30118e8984c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.210214 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cf38a30-f714-4a15-8be5-30118e8984c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5qth5\" (UID: \"0cf38a30-f714-4a15-8be5-30118e8984c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.215661 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0cf38a30-f714-4a15-8be5-30118e8984c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5qth5\" (UID: \"0cf38a30-f714-4a15-8be5-30118e8984c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.224776 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9wbt\" (UniqueName: \"kubernetes.io/projected/0cf38a30-f714-4a15-8be5-30118e8984c7-kube-api-access-t9wbt\") pod \"marketplace-operator-79b997595-5qth5\" (UID: \"0cf38a30-f714-4a15-8be5-30118e8984c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: E1004 08:20:48.230455 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7 is running failed: container process not found" containerID="fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 08:20:48 crc kubenswrapper[4969]: E1004 08:20:48.231747 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7 is running failed: container process not found" containerID="fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 08:20:48 crc kubenswrapper[4969]: E1004 08:20:48.232689 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7 is running failed: container process not found" containerID="fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 08:20:48 crc kubenswrapper[4969]: E1004 08:20:48.232728 4969 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-vpcfb" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerName="registry-server" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.403180 4969 generic.go:334] "Generic (PLEG): container finished" podID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerID="fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7" exitCode=0 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.403240 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpcfb" event={"ID":"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9","Type":"ContainerDied","Data":"fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7"} Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.418063 4969 generic.go:334] "Generic (PLEG): container finished" podID="d0a906da-f3d3-40cf-a714-2159253bb043" containerID="f50e9dce1265bd029868849547362f4dd15bfd5fdccb1f64cb08a3a8c76d7448" exitCode=0 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.418121 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" event={"ID":"d0a906da-f3d3-40cf-a714-2159253bb043","Type":"ContainerDied","Data":"f50e9dce1265bd029868849547362f4dd15bfd5fdccb1f64cb08a3a8c76d7448"} Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.422356 4969 generic.go:334] "Generic (PLEG): container finished" podID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerID="326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d" exitCode=0 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.422401 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9nrp" event={"ID":"46ed6532-e8c7-43c3-b80e-aeba1c5e2273","Type":"ContainerDied","Data":"326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d"} Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.424083 4969 generic.go:334] "Generic (PLEG): container finished" podID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerID="f31f51c0d54a4c6486ee1f841f9a4c6cd7da38de8bb7c5c201ab3a5726b21a88" exitCode=0 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.424114 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlrqv" event={"ID":"03b29be6-fac6-418e-97f1-a9cdad90fbf0","Type":"ContainerDied","Data":"f31f51c0d54a4c6486ee1f841f9a4c6cd7da38de8bb7c5c201ab3a5726b21a88"} Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.427315 4969 generic.go:334] "Generic (PLEG): container finished" podID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerID="d22ac309ab7c2844bf78fb7bf3c6c082207fb5f41cbfd59c401360dba1bf8fc3" exitCode=0 Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.427379 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxq79" event={"ID":"f0a988d3-5ee9-4833-97c8-05796123ff9a","Type":"ContainerDied","Data":"d22ac309ab7c2844bf78fb7bf3c6c082207fb5f41cbfd59c401360dba1bf8fc3"} Oct 04 08:20:48 crc kubenswrapper[4969]: E1004 08:20:48.481728 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d is running failed: container process not found" containerID="326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 08:20:48 crc kubenswrapper[4969]: E1004 08:20:48.482376 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d is running failed: container process not found" containerID="326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 08:20:48 crc kubenswrapper[4969]: E1004 08:20:48.482635 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d is running failed: container process not found" containerID="326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 08:20:48 crc kubenswrapper[4969]: E1004 08:20:48.482660 4969 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-b9nrp" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerName="registry-server" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.498839 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.505111 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.508818 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.519041 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.546928 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.570229 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.618895 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-catalog-content\") pod \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.618972 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-catalog-content\") pod \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.619027 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp6g7\" (UniqueName: \"kubernetes.io/projected/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-kube-api-access-lp6g7\") pod \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.619052 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhgqn\" (UniqueName: \"kubernetes.io/projected/03b29be6-fac6-418e-97f1-a9cdad90fbf0-kube-api-access-rhgqn\") pod \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.619087 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-utilities\") pod \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\" (UID: \"46ed6532-e8c7-43c3-b80e-aeba1c5e2273\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.619115 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-utilities\") pod \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\" (UID: \"03b29be6-fac6-418e-97f1-a9cdad90fbf0\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.620314 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-utilities" (OuterVolumeSpecName: "utilities") pod "03b29be6-fac6-418e-97f1-a9cdad90fbf0" (UID: "03b29be6-fac6-418e-97f1-a9cdad90fbf0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.623246 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-utilities" (OuterVolumeSpecName: "utilities") pod "46ed6532-e8c7-43c3-b80e-aeba1c5e2273" (UID: "46ed6532-e8c7-43c3-b80e-aeba1c5e2273"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.625948 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03b29be6-fac6-418e-97f1-a9cdad90fbf0-kube-api-access-rhgqn" (OuterVolumeSpecName: "kube-api-access-rhgqn") pod "03b29be6-fac6-418e-97f1-a9cdad90fbf0" (UID: "03b29be6-fac6-418e-97f1-a9cdad90fbf0"). InnerVolumeSpecName "kube-api-access-rhgqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.626038 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-kube-api-access-lp6g7" (OuterVolumeSpecName: "kube-api-access-lp6g7") pod "46ed6532-e8c7-43c3-b80e-aeba1c5e2273" (UID: "46ed6532-e8c7-43c3-b80e-aeba1c5e2273"). InnerVolumeSpecName "kube-api-access-lp6g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.650805 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03b29be6-fac6-418e-97f1-a9cdad90fbf0" (UID: "03b29be6-fac6-418e-97f1-a9cdad90fbf0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.689384 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46ed6532-e8c7-43c3-b80e-aeba1c5e2273" (UID: "46ed6532-e8c7-43c3-b80e-aeba1c5e2273"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.720121 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-utilities\") pod \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.720159 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-catalog-content\") pod \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.721343 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-utilities" (OuterVolumeSpecName: "utilities") pod "55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" (UID: "55366ee0-2720-4e4a-83e3-c4fb9be1f6a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.728518 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzzwv\" (UniqueName: \"kubernetes.io/projected/f0a988d3-5ee9-4833-97c8-05796123ff9a-kube-api-access-vzzwv\") pod \"f0a988d3-5ee9-4833-97c8-05796123ff9a\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.728610 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-trusted-ca\") pod \"d0a906da-f3d3-40cf-a714-2159253bb043\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.728642 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8kxc\" (UniqueName: \"kubernetes.io/projected/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-kube-api-access-b8kxc\") pod \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\" (UID: \"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.728682 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-operator-metrics\") pod \"d0a906da-f3d3-40cf-a714-2159253bb043\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.728707 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-catalog-content\") pod \"f0a988d3-5ee9-4833-97c8-05796123ff9a\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.728747 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-utilities\") pod \"f0a988d3-5ee9-4833-97c8-05796123ff9a\" (UID: \"f0a988d3-5ee9-4833-97c8-05796123ff9a\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.728774 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn586\" (UniqueName: \"kubernetes.io/projected/d0a906da-f3d3-40cf-a714-2159253bb043-kube-api-access-nn586\") pod \"d0a906da-f3d3-40cf-a714-2159253bb043\" (UID: \"d0a906da-f3d3-40cf-a714-2159253bb043\") " Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.729330 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp6g7\" (UniqueName: \"kubernetes.io/projected/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-kube-api-access-lp6g7\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.729349 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhgqn\" (UniqueName: \"kubernetes.io/projected/03b29be6-fac6-418e-97f1-a9cdad90fbf0-kube-api-access-rhgqn\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.729363 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.729375 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.729388 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.729399 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b29be6-fac6-418e-97f1-a9cdad90fbf0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.729410 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ed6532-e8c7-43c3-b80e-aeba1c5e2273-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.729558 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-utilities" (OuterVolumeSpecName: "utilities") pod "f0a988d3-5ee9-4833-97c8-05796123ff9a" (UID: "f0a988d3-5ee9-4833-97c8-05796123ff9a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.729741 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "d0a906da-f3d3-40cf-a714-2159253bb043" (UID: "d0a906da-f3d3-40cf-a714-2159253bb043"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.731960 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0a988d3-5ee9-4833-97c8-05796123ff9a-kube-api-access-vzzwv" (OuterVolumeSpecName: "kube-api-access-vzzwv") pod "f0a988d3-5ee9-4833-97c8-05796123ff9a" (UID: "f0a988d3-5ee9-4833-97c8-05796123ff9a"). InnerVolumeSpecName "kube-api-access-vzzwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.732675 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0a906da-f3d3-40cf-a714-2159253bb043-kube-api-access-nn586" (OuterVolumeSpecName: "kube-api-access-nn586") pod "d0a906da-f3d3-40cf-a714-2159253bb043" (UID: "d0a906da-f3d3-40cf-a714-2159253bb043"). InnerVolumeSpecName "kube-api-access-nn586". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.732816 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "d0a906da-f3d3-40cf-a714-2159253bb043" (UID: "d0a906da-f3d3-40cf-a714-2159253bb043"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.732858 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-kube-api-access-b8kxc" (OuterVolumeSpecName: "kube-api-access-b8kxc") pod "55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" (UID: "55366ee0-2720-4e4a-83e3-c4fb9be1f6a9"). InnerVolumeSpecName "kube-api-access-b8kxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.770825 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" (UID: "55366ee0-2720-4e4a-83e3-c4fb9be1f6a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.815133 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0a988d3-5ee9-4833-97c8-05796123ff9a" (UID: "f0a988d3-5ee9-4833-97c8-05796123ff9a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.830151 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.830190 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzzwv\" (UniqueName: \"kubernetes.io/projected/f0a988d3-5ee9-4833-97c8-05796123ff9a-kube-api-access-vzzwv\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.830205 4969 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.830221 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8kxc\" (UniqueName: \"kubernetes.io/projected/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9-kube-api-access-b8kxc\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.830232 4969 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d0a906da-f3d3-40cf-a714-2159253bb043-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.830245 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.830258 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a988d3-5ee9-4833-97c8-05796123ff9a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.830272 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn586\" (UniqueName: \"kubernetes.io/projected/d0a906da-f3d3-40cf-a714-2159253bb043-kube-api-access-nn586\") on node \"crc\" DevicePath \"\"" Oct 04 08:20:48 crc kubenswrapper[4969]: I1004 08:20:48.931600 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5qth5"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.441376 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlrqv" event={"ID":"03b29be6-fac6-418e-97f1-a9cdad90fbf0","Type":"ContainerDied","Data":"ef2cbb055512836775d751923f304b924bfcc4b3ea4fc258ae6456b4f6f9f5a6"} Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.441733 4969 scope.go:117] "RemoveContainer" containerID="f31f51c0d54a4c6486ee1f841f9a4c6cd7da38de8bb7c5c201ab3a5726b21a88" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.441460 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlrqv" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.444343 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" event={"ID":"0cf38a30-f714-4a15-8be5-30118e8984c7","Type":"ContainerStarted","Data":"17834281b10f1af331d3c2caad3a65c0f5079af16465e7dccc8cd67fb6657985"} Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.444392 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" event={"ID":"0cf38a30-f714-4a15-8be5-30118e8984c7","Type":"ContainerStarted","Data":"6b2d7b9c2827723174687ca9a31c474cd10dcc265c5cc4ad5af106a84ddc7e67"} Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.444952 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.446353 4969 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5qth5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.446403 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" podUID="0cf38a30-f714-4a15-8be5-30118e8984c7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.448966 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxq79" event={"ID":"f0a988d3-5ee9-4833-97c8-05796123ff9a","Type":"ContainerDied","Data":"729be080ffac341b25560a4365ab215d476ac6a7835a379d7ee60d06eb438f60"} Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.449006 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxq79" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.455862 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpcfb" event={"ID":"55366ee0-2720-4e4a-83e3-c4fb9be1f6a9","Type":"ContainerDied","Data":"5026eb0112bcef88d6fe02e27fde020a49d777530073c6b5eb7ff2fcd9c055f8"} Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.455874 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vpcfb" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.458706 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlrqv"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.460298 4969 scope.go:117] "RemoveContainer" containerID="e48e2341c4cb2495c33417a8b410b2bddaf7b030fe3cc5ce43f693bcdf8b5786" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.467333 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlrqv"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.471836 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" event={"ID":"d0a906da-f3d3-40cf-a714-2159253bb043","Type":"ContainerDied","Data":"bfa614b3b46865acaad3e136a026af4269c30def36d20a5a19ce737b9f25dfab"} Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.472011 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7tnxs" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.477302 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9nrp" event={"ID":"46ed6532-e8c7-43c3-b80e-aeba1c5e2273","Type":"ContainerDied","Data":"cd545504a5ec71ba8ff615bc45354d2ac0741855887563900fac792c681dd28e"} Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.477405 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9nrp" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.478609 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" podStartSLOduration=1.478592011 podStartE2EDuration="1.478592011s" podCreationTimestamp="2025-10-04 08:20:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:20:49.478280374 +0000 UTC m=+277.232549188" watchObservedRunningTime="2025-10-04 08:20:49.478592011 +0000 UTC m=+277.232860825" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.494068 4969 scope.go:117] "RemoveContainer" containerID="be80e8d8059750c4949b13ea155aa7a6845fd994b5bd786b661109315328fabc" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.501395 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vpcfb"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.510783 4969 scope.go:117] "RemoveContainer" containerID="d22ac309ab7c2844bf78fb7bf3c6c082207fb5f41cbfd59c401360dba1bf8fc3" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.511312 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vpcfb"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.522576 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7tnxs"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.530056 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7tnxs"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.534706 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxq79"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.536550 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mxq79"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.545897 4969 scope.go:117] "RemoveContainer" containerID="f2b87ed3e0ff0a5d7c383cb4d6f354e8f49109c9ad588fc33985e95173c7c5d8" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.550139 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9nrp"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.554658 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b9nrp"] Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.569656 4969 scope.go:117] "RemoveContainer" containerID="f52a5bbb0c745547ca030f2fcbd2a7057b9eae68db08f6c3ecc1fb44ffa2031e" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.586799 4969 scope.go:117] "RemoveContainer" containerID="fba9ac59e5766d0447a0f549fd6fbe14cd6a0078153631c8783bc584843b0ca7" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.605590 4969 scope.go:117] "RemoveContainer" containerID="620acb639ea37cc8ddb372215f9591b088a11f8ac91abfe89fa97c4981278464" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.619200 4969 scope.go:117] "RemoveContainer" containerID="82be18e59c05e559f6ddb241d8530a4817aea0513a0d0287256b93d7dd615780" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.635296 4969 scope.go:117] "RemoveContainer" containerID="f50e9dce1265bd029868849547362f4dd15bfd5fdccb1f64cb08a3a8c76d7448" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.649576 4969 scope.go:117] "RemoveContainer" containerID="326a8af2776c4550072ea11f4bc8b84ed88a5986d074e91a21747091a5f3ea3d" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.664501 4969 scope.go:117] "RemoveContainer" containerID="e361fa8e23ae2122bb753bd71935ceef9f2582e0c453d4357914671cb4f8ef87" Oct 04 08:20:49 crc kubenswrapper[4969]: I1004 08:20:49.680048 4969 scope.go:117] "RemoveContainer" containerID="6389c63785aca13aac72855d1369bd2d3a94697d57c6138bb11ba42dbb5a1248" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.083327 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xrjgj"] Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.083930 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.083947 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.083965 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a906da-f3d3-40cf-a714-2159253bb043" containerName="marketplace-operator" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.083974 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a906da-f3d3-40cf-a714-2159253bb043" containerName="marketplace-operator" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.083985 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerName="extract-utilities" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.083994 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerName="extract-utilities" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084003 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerName="extract-utilities" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084010 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerName="extract-utilities" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084021 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084029 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084040 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerName="extract-utilities" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084048 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerName="extract-utilities" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084058 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerName="extract-content" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084065 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerName="extract-content" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084079 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084088 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084104 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerName="extract-content" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084112 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerName="extract-content" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084123 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerName="extract-content" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084133 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerName="extract-content" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084146 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerName="extract-content" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084154 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerName="extract-content" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084166 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerName="extract-utilities" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084175 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerName="extract-utilities" Oct 04 08:20:50 crc kubenswrapper[4969]: E1004 08:20:50.084186 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084193 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084302 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084321 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0a906da-f3d3-40cf-a714-2159253bb043" containerName="marketplace-operator" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084336 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084347 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.084363 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" containerName="registry-server" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.085247 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.088084 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.091049 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrjgj"] Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.155973 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34db4256-784d-48a4-ae67-47e33b1c0e9f-utilities\") pod \"redhat-marketplace-xrjgj\" (UID: \"34db4256-784d-48a4-ae67-47e33b1c0e9f\") " pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.156103 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34db4256-784d-48a4-ae67-47e33b1c0e9f-catalog-content\") pod \"redhat-marketplace-xrjgj\" (UID: \"34db4256-784d-48a4-ae67-47e33b1c0e9f\") " pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.156213 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hh7s\" (UniqueName: \"kubernetes.io/projected/34db4256-784d-48a4-ae67-47e33b1c0e9f-kube-api-access-2hh7s\") pod \"redhat-marketplace-xrjgj\" (UID: \"34db4256-784d-48a4-ae67-47e33b1c0e9f\") " pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.256960 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34db4256-784d-48a4-ae67-47e33b1c0e9f-utilities\") pod \"redhat-marketplace-xrjgj\" (UID: \"34db4256-784d-48a4-ae67-47e33b1c0e9f\") " pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.257239 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34db4256-784d-48a4-ae67-47e33b1c0e9f-catalog-content\") pod \"redhat-marketplace-xrjgj\" (UID: \"34db4256-784d-48a4-ae67-47e33b1c0e9f\") " pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.257353 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hh7s\" (UniqueName: \"kubernetes.io/projected/34db4256-784d-48a4-ae67-47e33b1c0e9f-kube-api-access-2hh7s\") pod \"redhat-marketplace-xrjgj\" (UID: \"34db4256-784d-48a4-ae67-47e33b1c0e9f\") " pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.257644 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34db4256-784d-48a4-ae67-47e33b1c0e9f-utilities\") pod \"redhat-marketplace-xrjgj\" (UID: \"34db4256-784d-48a4-ae67-47e33b1c0e9f\") " pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.257748 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34db4256-784d-48a4-ae67-47e33b1c0e9f-catalog-content\") pod \"redhat-marketplace-xrjgj\" (UID: \"34db4256-784d-48a4-ae67-47e33b1c0e9f\") " pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.279805 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hh7s\" (UniqueName: \"kubernetes.io/projected/34db4256-784d-48a4-ae67-47e33b1c0e9f-kube-api-access-2hh7s\") pod \"redhat-marketplace-xrjgj\" (UID: \"34db4256-784d-48a4-ae67-47e33b1c0e9f\") " pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.401675 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.499244 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5qth5" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.672385 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lwlvl"] Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.674480 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.677372 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.690210 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwlvl"] Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.763716 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d378dd74-7b3b-4e78-9815-f75e40f48d1f-utilities\") pod \"certified-operators-lwlvl\" (UID: \"d378dd74-7b3b-4e78-9815-f75e40f48d1f\") " pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.763822 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d378dd74-7b3b-4e78-9815-f75e40f48d1f-catalog-content\") pod \"certified-operators-lwlvl\" (UID: \"d378dd74-7b3b-4e78-9815-f75e40f48d1f\") " pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.763861 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-668vh\" (UniqueName: \"kubernetes.io/projected/d378dd74-7b3b-4e78-9815-f75e40f48d1f-kube-api-access-668vh\") pod \"certified-operators-lwlvl\" (UID: \"d378dd74-7b3b-4e78-9815-f75e40f48d1f\") " pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.826730 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrjgj"] Oct 04 08:20:50 crc kubenswrapper[4969]: W1004 08:20:50.830056 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34db4256_784d_48a4_ae67_47e33b1c0e9f.slice/crio-38f40e044a5d99dea3f542bae4fdd838eecd201e29b55affa15bd4eae93d66fe WatchSource:0}: Error finding container 38f40e044a5d99dea3f542bae4fdd838eecd201e29b55affa15bd4eae93d66fe: Status 404 returned error can't find the container with id 38f40e044a5d99dea3f542bae4fdd838eecd201e29b55affa15bd4eae93d66fe Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.865631 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d378dd74-7b3b-4e78-9815-f75e40f48d1f-utilities\") pod \"certified-operators-lwlvl\" (UID: \"d378dd74-7b3b-4e78-9815-f75e40f48d1f\") " pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.865699 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d378dd74-7b3b-4e78-9815-f75e40f48d1f-catalog-content\") pod \"certified-operators-lwlvl\" (UID: \"d378dd74-7b3b-4e78-9815-f75e40f48d1f\") " pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.865734 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-668vh\" (UniqueName: \"kubernetes.io/projected/d378dd74-7b3b-4e78-9815-f75e40f48d1f-kube-api-access-668vh\") pod \"certified-operators-lwlvl\" (UID: \"d378dd74-7b3b-4e78-9815-f75e40f48d1f\") " pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.866329 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d378dd74-7b3b-4e78-9815-f75e40f48d1f-utilities\") pod \"certified-operators-lwlvl\" (UID: \"d378dd74-7b3b-4e78-9815-f75e40f48d1f\") " pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.866339 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d378dd74-7b3b-4e78-9815-f75e40f48d1f-catalog-content\") pod \"certified-operators-lwlvl\" (UID: \"d378dd74-7b3b-4e78-9815-f75e40f48d1f\") " pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.889333 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-668vh\" (UniqueName: \"kubernetes.io/projected/d378dd74-7b3b-4e78-9815-f75e40f48d1f-kube-api-access-668vh\") pod \"certified-operators-lwlvl\" (UID: \"d378dd74-7b3b-4e78-9815-f75e40f48d1f\") " pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:50 crc kubenswrapper[4969]: I1004 08:20:50.998052 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.063752 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03b29be6-fac6-418e-97f1-a9cdad90fbf0" path="/var/lib/kubelet/pods/03b29be6-fac6-418e-97f1-a9cdad90fbf0/volumes" Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.064506 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46ed6532-e8c7-43c3-b80e-aeba1c5e2273" path="/var/lib/kubelet/pods/46ed6532-e8c7-43c3-b80e-aeba1c5e2273/volumes" Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.065186 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55366ee0-2720-4e4a-83e3-c4fb9be1f6a9" path="/var/lib/kubelet/pods/55366ee0-2720-4e4a-83e3-c4fb9be1f6a9/volumes" Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.066643 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0a906da-f3d3-40cf-a714-2159253bb043" path="/var/lib/kubelet/pods/d0a906da-f3d3-40cf-a714-2159253bb043/volumes" Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.067178 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0a988d3-5ee9-4833-97c8-05796123ff9a" path="/var/lib/kubelet/pods/f0a988d3-5ee9-4833-97c8-05796123ff9a/volumes" Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.242485 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwlvl"] Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.504390 4969 generic.go:334] "Generic (PLEG): container finished" podID="34db4256-784d-48a4-ae67-47e33b1c0e9f" containerID="288cf3831492bd2f4b40fdb09d8b4fbe4540cae90262c03b7b2a1e0d0c289f48" exitCode=0 Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.504455 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrjgj" event={"ID":"34db4256-784d-48a4-ae67-47e33b1c0e9f","Type":"ContainerDied","Data":"288cf3831492bd2f4b40fdb09d8b4fbe4540cae90262c03b7b2a1e0d0c289f48"} Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.504499 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrjgj" event={"ID":"34db4256-784d-48a4-ae67-47e33b1c0e9f","Type":"ContainerStarted","Data":"38f40e044a5d99dea3f542bae4fdd838eecd201e29b55affa15bd4eae93d66fe"} Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.510268 4969 generic.go:334] "Generic (PLEG): container finished" podID="d378dd74-7b3b-4e78-9815-f75e40f48d1f" containerID="8745447cb0ff88fceb541bf7f8241b271e875a8b6e5ec4832107ac430e20870c" exitCode=0 Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.510335 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwlvl" event={"ID":"d378dd74-7b3b-4e78-9815-f75e40f48d1f","Type":"ContainerDied","Data":"8745447cb0ff88fceb541bf7f8241b271e875a8b6e5ec4832107ac430e20870c"} Oct 04 08:20:51 crc kubenswrapper[4969]: I1004 08:20:51.510407 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwlvl" event={"ID":"d378dd74-7b3b-4e78-9815-f75e40f48d1f","Type":"ContainerStarted","Data":"8c627f70660bd45d46e37cc9a039cd6458f3773aebe69592540f9ae82c1e12c7"} Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.485299 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p7tnv"] Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.487090 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.489191 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.515068 4969 generic.go:334] "Generic (PLEG): container finished" podID="34db4256-784d-48a4-ae67-47e33b1c0e9f" containerID="5c876c6fb8f899a0b73923060be9724d17432c51033ce593a2ee2b00f8ea0c1d" exitCode=0 Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.515115 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrjgj" event={"ID":"34db4256-784d-48a4-ae67-47e33b1c0e9f","Type":"ContainerDied","Data":"5c876c6fb8f899a0b73923060be9724d17432c51033ce593a2ee2b00f8ea0c1d"} Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.533338 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p7tnv"] Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.587576 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a59cef3-965a-4244-89af-0eaa80eef618-catalog-content\") pod \"redhat-operators-p7tnv\" (UID: \"3a59cef3-965a-4244-89af-0eaa80eef618\") " pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.587668 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9dhz\" (UniqueName: \"kubernetes.io/projected/3a59cef3-965a-4244-89af-0eaa80eef618-kube-api-access-x9dhz\") pod \"redhat-operators-p7tnv\" (UID: \"3a59cef3-965a-4244-89af-0eaa80eef618\") " pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.587751 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a59cef3-965a-4244-89af-0eaa80eef618-utilities\") pod \"redhat-operators-p7tnv\" (UID: \"3a59cef3-965a-4244-89af-0eaa80eef618\") " pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.689497 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9dhz\" (UniqueName: \"kubernetes.io/projected/3a59cef3-965a-4244-89af-0eaa80eef618-kube-api-access-x9dhz\") pod \"redhat-operators-p7tnv\" (UID: \"3a59cef3-965a-4244-89af-0eaa80eef618\") " pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.689577 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a59cef3-965a-4244-89af-0eaa80eef618-utilities\") pod \"redhat-operators-p7tnv\" (UID: \"3a59cef3-965a-4244-89af-0eaa80eef618\") " pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.689690 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a59cef3-965a-4244-89af-0eaa80eef618-catalog-content\") pod \"redhat-operators-p7tnv\" (UID: \"3a59cef3-965a-4244-89af-0eaa80eef618\") " pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.690678 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a59cef3-965a-4244-89af-0eaa80eef618-utilities\") pod \"redhat-operators-p7tnv\" (UID: \"3a59cef3-965a-4244-89af-0eaa80eef618\") " pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.690707 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a59cef3-965a-4244-89af-0eaa80eef618-catalog-content\") pod \"redhat-operators-p7tnv\" (UID: \"3a59cef3-965a-4244-89af-0eaa80eef618\") " pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.716343 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9dhz\" (UniqueName: \"kubernetes.io/projected/3a59cef3-965a-4244-89af-0eaa80eef618-kube-api-access-x9dhz\") pod \"redhat-operators-p7tnv\" (UID: \"3a59cef3-965a-4244-89af-0eaa80eef618\") " pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:52 crc kubenswrapper[4969]: I1004 08:20:52.867538 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.098298 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5drz6"] Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.103354 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.119031 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.120188 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5drz6"] Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.201357 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9jgd\" (UniqueName: \"kubernetes.io/projected/d8276ca7-275d-4cff-8895-f868dab88c73-kube-api-access-z9jgd\") pod \"community-operators-5drz6\" (UID: \"d8276ca7-275d-4cff-8895-f868dab88c73\") " pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.201443 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8276ca7-275d-4cff-8895-f868dab88c73-utilities\") pod \"community-operators-5drz6\" (UID: \"d8276ca7-275d-4cff-8895-f868dab88c73\") " pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.201464 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8276ca7-275d-4cff-8895-f868dab88c73-catalog-content\") pod \"community-operators-5drz6\" (UID: \"d8276ca7-275d-4cff-8895-f868dab88c73\") " pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.302625 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9jgd\" (UniqueName: \"kubernetes.io/projected/d8276ca7-275d-4cff-8895-f868dab88c73-kube-api-access-z9jgd\") pod \"community-operators-5drz6\" (UID: \"d8276ca7-275d-4cff-8895-f868dab88c73\") " pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.302688 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8276ca7-275d-4cff-8895-f868dab88c73-catalog-content\") pod \"community-operators-5drz6\" (UID: \"d8276ca7-275d-4cff-8895-f868dab88c73\") " pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.302704 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8276ca7-275d-4cff-8895-f868dab88c73-utilities\") pod \"community-operators-5drz6\" (UID: \"d8276ca7-275d-4cff-8895-f868dab88c73\") " pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.303138 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8276ca7-275d-4cff-8895-f868dab88c73-utilities\") pod \"community-operators-5drz6\" (UID: \"d8276ca7-275d-4cff-8895-f868dab88c73\") " pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.303632 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8276ca7-275d-4cff-8895-f868dab88c73-catalog-content\") pod \"community-operators-5drz6\" (UID: \"d8276ca7-275d-4cff-8895-f868dab88c73\") " pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.315576 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p7tnv"] Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.321314 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9jgd\" (UniqueName: \"kubernetes.io/projected/d8276ca7-275d-4cff-8895-f868dab88c73-kube-api-access-z9jgd\") pod \"community-operators-5drz6\" (UID: \"d8276ca7-275d-4cff-8895-f868dab88c73\") " pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: W1004 08:20:53.321374 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a59cef3_965a_4244_89af_0eaa80eef618.slice/crio-aeb4e27ea163435a61cefb5e83b95dc367343dcf70ee5c47522dac15aae68acd WatchSource:0}: Error finding container aeb4e27ea163435a61cefb5e83b95dc367343dcf70ee5c47522dac15aae68acd: Status 404 returned error can't find the container with id aeb4e27ea163435a61cefb5e83b95dc367343dcf70ee5c47522dac15aae68acd Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.431052 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.526068 4969 generic.go:334] "Generic (PLEG): container finished" podID="3a59cef3-965a-4244-89af-0eaa80eef618" containerID="dd2320c0b3a6a9d66fd7bd0a2ddb1a7de99fbbf40537aafe63d14d3154020ca5" exitCode=0 Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.526146 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7tnv" event={"ID":"3a59cef3-965a-4244-89af-0eaa80eef618","Type":"ContainerDied","Data":"dd2320c0b3a6a9d66fd7bd0a2ddb1a7de99fbbf40537aafe63d14d3154020ca5"} Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.526404 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7tnv" event={"ID":"3a59cef3-965a-4244-89af-0eaa80eef618","Type":"ContainerStarted","Data":"aeb4e27ea163435a61cefb5e83b95dc367343dcf70ee5c47522dac15aae68acd"} Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.531998 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrjgj" event={"ID":"34db4256-784d-48a4-ae67-47e33b1c0e9f","Type":"ContainerStarted","Data":"1c0b4ceeb0b7303d797fce8ee06bfa6459035dfa51812fa200d2acf0987cdf68"} Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.537968 4969 generic.go:334] "Generic (PLEG): container finished" podID="d378dd74-7b3b-4e78-9815-f75e40f48d1f" containerID="506a463d0b06e551ee5a7346b5b4ce6a68da3b9769f5204b363dce8f4a8068b7" exitCode=0 Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.538017 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwlvl" event={"ID":"d378dd74-7b3b-4e78-9815-f75e40f48d1f","Type":"ContainerDied","Data":"506a463d0b06e551ee5a7346b5b4ce6a68da3b9769f5204b363dce8f4a8068b7"} Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.583161 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xrjgj" podStartSLOduration=2.109552855 podStartE2EDuration="3.583139949s" podCreationTimestamp="2025-10-04 08:20:50 +0000 UTC" firstStartedPulling="2025-10-04 08:20:51.507614996 +0000 UTC m=+279.261883810" lastFinishedPulling="2025-10-04 08:20:52.98120209 +0000 UTC m=+280.735470904" observedRunningTime="2025-10-04 08:20:53.577433727 +0000 UTC m=+281.331702541" watchObservedRunningTime="2025-10-04 08:20:53.583139949 +0000 UTC m=+281.337408763" Oct 04 08:20:53 crc kubenswrapper[4969]: I1004 08:20:53.637351 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5drz6"] Oct 04 08:20:53 crc kubenswrapper[4969]: W1004 08:20:53.649620 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8276ca7_275d_4cff_8895_f868dab88c73.slice/crio-fd4f59c1e133288f851ba8dfc4476403a25279b5a2a0880bc333f58baeef539c WatchSource:0}: Error finding container fd4f59c1e133288f851ba8dfc4476403a25279b5a2a0880bc333f58baeef539c: Status 404 returned error can't find the container with id fd4f59c1e133288f851ba8dfc4476403a25279b5a2a0880bc333f58baeef539c Oct 04 08:20:54 crc kubenswrapper[4969]: I1004 08:20:54.545830 4969 generic.go:334] "Generic (PLEG): container finished" podID="d8276ca7-275d-4cff-8895-f868dab88c73" containerID="7290e9b10b67d19ba9c40f0d8f8dee7b1832d67a864e571c41843c22aea63cc5" exitCode=0 Oct 04 08:20:54 crc kubenswrapper[4969]: I1004 08:20:54.545877 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5drz6" event={"ID":"d8276ca7-275d-4cff-8895-f868dab88c73","Type":"ContainerDied","Data":"7290e9b10b67d19ba9c40f0d8f8dee7b1832d67a864e571c41843c22aea63cc5"} Oct 04 08:20:54 crc kubenswrapper[4969]: I1004 08:20:54.546289 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5drz6" event={"ID":"d8276ca7-275d-4cff-8895-f868dab88c73","Type":"ContainerStarted","Data":"fd4f59c1e133288f851ba8dfc4476403a25279b5a2a0880bc333f58baeef539c"} Oct 04 08:20:54 crc kubenswrapper[4969]: I1004 08:20:54.548594 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwlvl" event={"ID":"d378dd74-7b3b-4e78-9815-f75e40f48d1f","Type":"ContainerStarted","Data":"52b747c05e94397e73e16de2423ab66fe1f293fd2ddfaf44b5fa7f7f0e969775"} Oct 04 08:20:54 crc kubenswrapper[4969]: I1004 08:20:54.553948 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7tnv" event={"ID":"3a59cef3-965a-4244-89af-0eaa80eef618","Type":"ContainerStarted","Data":"5b1663535c1c09b41cd4f5a47de5017498a25cdd1e42aaa70cb5aaec2b79ac61"} Oct 04 08:20:54 crc kubenswrapper[4969]: I1004 08:20:54.580356 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lwlvl" podStartSLOduration=2.056751559 podStartE2EDuration="4.58033428s" podCreationTimestamp="2025-10-04 08:20:50 +0000 UTC" firstStartedPulling="2025-10-04 08:20:51.512018115 +0000 UTC m=+279.266286929" lastFinishedPulling="2025-10-04 08:20:54.035600806 +0000 UTC m=+281.789869650" observedRunningTime="2025-10-04 08:20:54.57634206 +0000 UTC m=+282.330610864" watchObservedRunningTime="2025-10-04 08:20:54.58033428 +0000 UTC m=+282.334603104" Oct 04 08:20:55 crc kubenswrapper[4969]: I1004 08:20:55.561147 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5drz6" event={"ID":"d8276ca7-275d-4cff-8895-f868dab88c73","Type":"ContainerStarted","Data":"6d2823ba4788a5a3bd7ca4c66826f1d681e115c7e73fae9d280280a1ee191020"} Oct 04 08:20:55 crc kubenswrapper[4969]: I1004 08:20:55.563390 4969 generic.go:334] "Generic (PLEG): container finished" podID="3a59cef3-965a-4244-89af-0eaa80eef618" containerID="5b1663535c1c09b41cd4f5a47de5017498a25cdd1e42aaa70cb5aaec2b79ac61" exitCode=0 Oct 04 08:20:55 crc kubenswrapper[4969]: I1004 08:20:55.563455 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7tnv" event={"ID":"3a59cef3-965a-4244-89af-0eaa80eef618","Type":"ContainerDied","Data":"5b1663535c1c09b41cd4f5a47de5017498a25cdd1e42aaa70cb5aaec2b79ac61"} Oct 04 08:20:56 crc kubenswrapper[4969]: I1004 08:20:56.568099 4969 generic.go:334] "Generic (PLEG): container finished" podID="d8276ca7-275d-4cff-8895-f868dab88c73" containerID="6d2823ba4788a5a3bd7ca4c66826f1d681e115c7e73fae9d280280a1ee191020" exitCode=0 Oct 04 08:20:56 crc kubenswrapper[4969]: I1004 08:20:56.568431 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5drz6" event={"ID":"d8276ca7-275d-4cff-8895-f868dab88c73","Type":"ContainerDied","Data":"6d2823ba4788a5a3bd7ca4c66826f1d681e115c7e73fae9d280280a1ee191020"} Oct 04 08:20:57 crc kubenswrapper[4969]: I1004 08:20:57.576963 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5drz6" event={"ID":"d8276ca7-275d-4cff-8895-f868dab88c73","Type":"ContainerStarted","Data":"1a806a590237e4b7a3e7db8f9028a4f0e7d50ca824258dfd83fc6c3ae4199ae7"} Oct 04 08:20:57 crc kubenswrapper[4969]: I1004 08:20:57.580404 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7tnv" event={"ID":"3a59cef3-965a-4244-89af-0eaa80eef618","Type":"ContainerStarted","Data":"f351d798955c40ab9a256383b2f9095fbc026e751ffba28907ded06f541bda73"} Oct 04 08:20:57 crc kubenswrapper[4969]: I1004 08:20:57.622616 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5drz6" podStartSLOduration=2.161244435 podStartE2EDuration="4.622588335s" podCreationTimestamp="2025-10-04 08:20:53 +0000 UTC" firstStartedPulling="2025-10-04 08:20:54.548167169 +0000 UTC m=+282.302435983" lastFinishedPulling="2025-10-04 08:20:57.009511029 +0000 UTC m=+284.763779883" observedRunningTime="2025-10-04 08:20:57.600769591 +0000 UTC m=+285.355038425" watchObservedRunningTime="2025-10-04 08:20:57.622588335 +0000 UTC m=+285.376857159" Oct 04 08:21:00 crc kubenswrapper[4969]: I1004 08:21:00.402355 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:21:00 crc kubenswrapper[4969]: I1004 08:21:00.402623 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:21:00 crc kubenswrapper[4969]: I1004 08:21:00.459121 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:21:00 crc kubenswrapper[4969]: I1004 08:21:00.481184 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p7tnv" podStartSLOduration=6.0639450440000005 podStartE2EDuration="8.481161775s" podCreationTimestamp="2025-10-04 08:20:52 +0000 UTC" firstStartedPulling="2025-10-04 08:20:53.527499013 +0000 UTC m=+281.281767827" lastFinishedPulling="2025-10-04 08:20:55.944715744 +0000 UTC m=+283.698984558" observedRunningTime="2025-10-04 08:20:57.621016356 +0000 UTC m=+285.375285200" watchObservedRunningTime="2025-10-04 08:21:00.481161775 +0000 UTC m=+288.235430589" Oct 04 08:21:00 crc kubenswrapper[4969]: I1004 08:21:00.656391 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xrjgj" Oct 04 08:21:00 crc kubenswrapper[4969]: I1004 08:21:00.998782 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:21:00 crc kubenswrapper[4969]: I1004 08:21:00.999081 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:21:01 crc kubenswrapper[4969]: I1004 08:21:01.065482 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:21:01 crc kubenswrapper[4969]: I1004 08:21:01.640368 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lwlvl" Oct 04 08:21:02 crc kubenswrapper[4969]: I1004 08:21:02.868557 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:21:02 crc kubenswrapper[4969]: I1004 08:21:02.868914 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:21:02 crc kubenswrapper[4969]: I1004 08:21:02.908228 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:21:03 crc kubenswrapper[4969]: I1004 08:21:03.432139 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:21:03 crc kubenswrapper[4969]: I1004 08:21:03.432600 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:21:03 crc kubenswrapper[4969]: I1004 08:21:03.481959 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:21:03 crc kubenswrapper[4969]: I1004 08:21:03.666327 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5drz6" Oct 04 08:21:03 crc kubenswrapper[4969]: I1004 08:21:03.669299 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p7tnv" Oct 04 08:21:49 crc kubenswrapper[4969]: I1004 08:21:49.667139 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:21:49 crc kubenswrapper[4969]: I1004 08:21:49.667746 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:22:19 crc kubenswrapper[4969]: I1004 08:22:19.667131 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:22:19 crc kubenswrapper[4969]: I1004 08:22:19.667931 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:22:49 crc kubenswrapper[4969]: I1004 08:22:49.666762 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:22:49 crc kubenswrapper[4969]: I1004 08:22:49.667298 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:22:49 crc kubenswrapper[4969]: I1004 08:22:49.667347 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:22:49 crc kubenswrapper[4969]: I1004 08:22:49.668055 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d265f04fbe19dd6f5ad0bc22c79d5d325f5680917d64f171975ad5c10938d825"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:22:49 crc kubenswrapper[4969]: I1004 08:22:49.668142 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://d265f04fbe19dd6f5ad0bc22c79d5d325f5680917d64f171975ad5c10938d825" gracePeriod=600 Oct 04 08:22:50 crc kubenswrapper[4969]: I1004 08:22:50.338277 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="d265f04fbe19dd6f5ad0bc22c79d5d325f5680917d64f171975ad5c10938d825" exitCode=0 Oct 04 08:22:50 crc kubenswrapper[4969]: I1004 08:22:50.338315 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"d265f04fbe19dd6f5ad0bc22c79d5d325f5680917d64f171975ad5c10938d825"} Oct 04 08:22:50 crc kubenswrapper[4969]: I1004 08:22:50.338663 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"89de65150d90f75568f74a82023f8dca9943d3891a3087d9e41cf1dc1b47d259"} Oct 04 08:22:50 crc kubenswrapper[4969]: I1004 08:22:50.338684 4969 scope.go:117] "RemoveContainer" containerID="01c904d3704e27fa97a787729e39cea7a94c7a4be331c8dbf91079b70f0e7452" Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.840009 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6lb59"] Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.841568 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.857540 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6lb59"] Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.984031 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b748b05-8935-4381-9cb1-8a3128764de5-bound-sa-token\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.984100 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b748b05-8935-4381-9cb1-8a3128764de5-trusted-ca\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.984124 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b748b05-8935-4381-9cb1-8a3128764de5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.984167 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl5xm\" (UniqueName: \"kubernetes.io/projected/8b748b05-8935-4381-9cb1-8a3128764de5-kube-api-access-rl5xm\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.984200 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b748b05-8935-4381-9cb1-8a3128764de5-registry-certificates\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.984228 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.984253 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b748b05-8935-4381-9cb1-8a3128764de5-registry-tls\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:13 crc kubenswrapper[4969]: I1004 08:23:13.984272 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b748b05-8935-4381-9cb1-8a3128764de5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.008249 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.085364 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b748b05-8935-4381-9cb1-8a3128764de5-bound-sa-token\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.085453 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b748b05-8935-4381-9cb1-8a3128764de5-trusted-ca\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.085474 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b748b05-8935-4381-9cb1-8a3128764de5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.085514 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl5xm\" (UniqueName: \"kubernetes.io/projected/8b748b05-8935-4381-9cb1-8a3128764de5-kube-api-access-rl5xm\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.085553 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b748b05-8935-4381-9cb1-8a3128764de5-registry-certificates\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.085584 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b748b05-8935-4381-9cb1-8a3128764de5-registry-tls\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.085610 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b748b05-8935-4381-9cb1-8a3128764de5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.086348 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b748b05-8935-4381-9cb1-8a3128764de5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.086954 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b748b05-8935-4381-9cb1-8a3128764de5-registry-certificates\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.087019 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b748b05-8935-4381-9cb1-8a3128764de5-trusted-ca\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.092186 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b748b05-8935-4381-9cb1-8a3128764de5-registry-tls\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.095113 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b748b05-8935-4381-9cb1-8a3128764de5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.117176 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b748b05-8935-4381-9cb1-8a3128764de5-bound-sa-token\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.117281 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl5xm\" (UniqueName: \"kubernetes.io/projected/8b748b05-8935-4381-9cb1-8a3128764de5-kube-api-access-rl5xm\") pod \"image-registry-66df7c8f76-6lb59\" (UID: \"8b748b05-8935-4381-9cb1-8a3128764de5\") " pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.158439 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.320352 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6lb59"] Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.502020 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" event={"ID":"8b748b05-8935-4381-9cb1-8a3128764de5","Type":"ContainerStarted","Data":"32b479ed06780b55b275f1cd7b1314da0022571833ed26680d4d9cb95b0d8776"} Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.502067 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" event={"ID":"8b748b05-8935-4381-9cb1-8a3128764de5","Type":"ContainerStarted","Data":"fe6a0586791a807c688f31114fbd9ef2e6dc0613cceaf7adf2565ed46ae6d52e"} Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.502186 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:14 crc kubenswrapper[4969]: I1004 08:23:14.521238 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" podStartSLOduration=1.521194907 podStartE2EDuration="1.521194907s" podCreationTimestamp="2025-10-04 08:23:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:23:14.517076695 +0000 UTC m=+422.271345509" watchObservedRunningTime="2025-10-04 08:23:14.521194907 +0000 UTC m=+422.275463721" Oct 04 08:23:34 crc kubenswrapper[4969]: I1004 08:23:34.164333 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-6lb59" Oct 04 08:23:34 crc kubenswrapper[4969]: I1004 08:23:34.244054 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cbwg8"] Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.299791 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" podUID="07fc39da-f348-4892-a8ab-df8200d861af" containerName="registry" containerID="cri-o://43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743" gracePeriod=30 Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.712623 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.802856 4969 generic.go:334] "Generic (PLEG): container finished" podID="07fc39da-f348-4892-a8ab-df8200d861af" containerID="43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743" exitCode=0 Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.802920 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" event={"ID":"07fc39da-f348-4892-a8ab-df8200d861af","Type":"ContainerDied","Data":"43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743"} Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.802948 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.802977 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cbwg8" event={"ID":"07fc39da-f348-4892-a8ab-df8200d861af","Type":"ContainerDied","Data":"0dd6a1d7c0c68e132d7d5846049143fbb1150bcba6149ad1932dfa8de677bbe2"} Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.803011 4969 scope.go:117] "RemoveContainer" containerID="43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.828410 4969 scope.go:117] "RemoveContainer" containerID="43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743" Oct 04 08:23:59 crc kubenswrapper[4969]: E1004 08:23:59.828982 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743\": container with ID starting with 43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743 not found: ID does not exist" containerID="43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.829037 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743"} err="failed to get container status \"43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743\": rpc error: code = NotFound desc = could not find container \"43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743\": container with ID starting with 43bf49fa6abc21c3bffa752afcfda6a8aed2267db015f4bf8ea358ae7cb25743 not found: ID does not exist" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.884553 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-registry-certificates\") pod \"07fc39da-f348-4892-a8ab-df8200d861af\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.884653 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t5dz\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-kube-api-access-5t5dz\") pod \"07fc39da-f348-4892-a8ab-df8200d861af\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.884721 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/07fc39da-f348-4892-a8ab-df8200d861af-ca-trust-extracted\") pod \"07fc39da-f348-4892-a8ab-df8200d861af\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.884789 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-bound-sa-token\") pod \"07fc39da-f348-4892-a8ab-df8200d861af\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.885031 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"07fc39da-f348-4892-a8ab-df8200d861af\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.885098 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-registry-tls\") pod \"07fc39da-f348-4892-a8ab-df8200d861af\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.885172 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-trusted-ca\") pod \"07fc39da-f348-4892-a8ab-df8200d861af\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.885220 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/07fc39da-f348-4892-a8ab-df8200d861af-installation-pull-secrets\") pod \"07fc39da-f348-4892-a8ab-df8200d861af\" (UID: \"07fc39da-f348-4892-a8ab-df8200d861af\") " Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.886039 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "07fc39da-f348-4892-a8ab-df8200d861af" (UID: "07fc39da-f348-4892-a8ab-df8200d861af"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.886091 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "07fc39da-f348-4892-a8ab-df8200d861af" (UID: "07fc39da-f348-4892-a8ab-df8200d861af"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.886891 4969 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.886935 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07fc39da-f348-4892-a8ab-df8200d861af-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.893569 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "07fc39da-f348-4892-a8ab-df8200d861af" (UID: "07fc39da-f348-4892-a8ab-df8200d861af"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.893867 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07fc39da-f348-4892-a8ab-df8200d861af-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "07fc39da-f348-4892-a8ab-df8200d861af" (UID: "07fc39da-f348-4892-a8ab-df8200d861af"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.897195 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-kube-api-access-5t5dz" (OuterVolumeSpecName: "kube-api-access-5t5dz") pod "07fc39da-f348-4892-a8ab-df8200d861af" (UID: "07fc39da-f348-4892-a8ab-df8200d861af"). InnerVolumeSpecName "kube-api-access-5t5dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.897493 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "07fc39da-f348-4892-a8ab-df8200d861af" (UID: "07fc39da-f348-4892-a8ab-df8200d861af"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.898344 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "07fc39da-f348-4892-a8ab-df8200d861af" (UID: "07fc39da-f348-4892-a8ab-df8200d861af"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.922998 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07fc39da-f348-4892-a8ab-df8200d861af-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "07fc39da-f348-4892-a8ab-df8200d861af" (UID: "07fc39da-f348-4892-a8ab-df8200d861af"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.988496 4969 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/07fc39da-f348-4892-a8ab-df8200d861af-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.988537 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t5dz\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-kube-api-access-5t5dz\") on node \"crc\" DevicePath \"\"" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.988552 4969 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/07fc39da-f348-4892-a8ab-df8200d861af-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.988564 4969 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 08:23:59 crc kubenswrapper[4969]: I1004 08:23:59.988576 4969 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/07fc39da-f348-4892-a8ab-df8200d861af-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:24:00 crc kubenswrapper[4969]: I1004 08:24:00.149552 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cbwg8"] Oct 04 08:24:00 crc kubenswrapper[4969]: I1004 08:24:00.155658 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cbwg8"] Oct 04 08:24:01 crc kubenswrapper[4969]: I1004 08:24:01.066557 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07fc39da-f348-4892-a8ab-df8200d861af" path="/var/lib/kubelet/pods/07fc39da-f348-4892-a8ab-df8200d861af/volumes" Oct 04 08:25:19 crc kubenswrapper[4969]: I1004 08:25:19.666765 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:25:19 crc kubenswrapper[4969]: I1004 08:25:19.667393 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:25:49 crc kubenswrapper[4969]: I1004 08:25:49.667053 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:25:49 crc kubenswrapper[4969]: I1004 08:25:49.667590 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:26:19 crc kubenswrapper[4969]: I1004 08:26:19.666736 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:26:19 crc kubenswrapper[4969]: I1004 08:26:19.667540 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:26:19 crc kubenswrapper[4969]: I1004 08:26:19.667612 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:26:19 crc kubenswrapper[4969]: I1004 08:26:19.668874 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89de65150d90f75568f74a82023f8dca9943d3891a3087d9e41cf1dc1b47d259"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:26:19 crc kubenswrapper[4969]: I1004 08:26:19.668985 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://89de65150d90f75568f74a82023f8dca9943d3891a3087d9e41cf1dc1b47d259" gracePeriod=600 Oct 04 08:26:20 crc kubenswrapper[4969]: I1004 08:26:20.714727 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="89de65150d90f75568f74a82023f8dca9943d3891a3087d9e41cf1dc1b47d259" exitCode=0 Oct 04 08:26:20 crc kubenswrapper[4969]: I1004 08:26:20.714787 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"89de65150d90f75568f74a82023f8dca9943d3891a3087d9e41cf1dc1b47d259"} Oct 04 08:26:20 crc kubenswrapper[4969]: I1004 08:26:20.715251 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"47fb93c2519799da0fd36fa2c011f7f4604debf093aa2814af4cd13eb886ba6f"} Oct 04 08:26:20 crc kubenswrapper[4969]: I1004 08:26:20.715298 4969 scope.go:117] "RemoveContainer" containerID="d265f04fbe19dd6f5ad0bc22c79d5d325f5680917d64f171975ad5c10938d825" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.708395 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-8s797"] Oct 04 08:26:49 crc kubenswrapper[4969]: E1004 08:26:49.709095 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07fc39da-f348-4892-a8ab-df8200d861af" containerName="registry" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.709109 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="07fc39da-f348-4892-a8ab-df8200d861af" containerName="registry" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.709218 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="07fc39da-f348-4892-a8ab-df8200d861af" containerName="registry" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.709573 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-8s797" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.715099 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.715201 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.716019 4969 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-7krfl" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.719158 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-8s797"] Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.721167 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w94hv"] Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.721883 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-w94hv" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.725044 4969 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-76pff" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.734264 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w94hv"] Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.745944 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rdw9w"] Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.746621 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.748480 4969 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-lrhdr" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.754120 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rdw9w"] Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.903792 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h74nv\" (UniqueName: \"kubernetes.io/projected/f37f93ad-8593-4bf2-bb4a-7d01e9ab273c-kube-api-access-h74nv\") pod \"cert-manager-webhook-5655c58dd6-rdw9w\" (UID: \"f37f93ad-8593-4bf2-bb4a-7d01e9ab273c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.903849 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8zwg\" (UniqueName: \"kubernetes.io/projected/21c43c6c-8aa0-4558-8e9e-ba34798b6c18-kube-api-access-g8zwg\") pod \"cert-manager-5b446d88c5-w94hv\" (UID: \"21c43c6c-8aa0-4558-8e9e-ba34798b6c18\") " pod="cert-manager/cert-manager-5b446d88c5-w94hv" Oct 04 08:26:49 crc kubenswrapper[4969]: I1004 08:26:49.903885 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcbtw\" (UniqueName: \"kubernetes.io/projected/daf3ed47-3e82-4a2a-ae17-d56aa124499e-kube-api-access-fcbtw\") pod \"cert-manager-cainjector-7f985d654d-8s797\" (UID: \"daf3ed47-3e82-4a2a-ae17-d56aa124499e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-8s797" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.005198 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8zwg\" (UniqueName: \"kubernetes.io/projected/21c43c6c-8aa0-4558-8e9e-ba34798b6c18-kube-api-access-g8zwg\") pod \"cert-manager-5b446d88c5-w94hv\" (UID: \"21c43c6c-8aa0-4558-8e9e-ba34798b6c18\") " pod="cert-manager/cert-manager-5b446d88c5-w94hv" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.005295 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcbtw\" (UniqueName: \"kubernetes.io/projected/daf3ed47-3e82-4a2a-ae17-d56aa124499e-kube-api-access-fcbtw\") pod \"cert-manager-cainjector-7f985d654d-8s797\" (UID: \"daf3ed47-3e82-4a2a-ae17-d56aa124499e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-8s797" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.005414 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h74nv\" (UniqueName: \"kubernetes.io/projected/f37f93ad-8593-4bf2-bb4a-7d01e9ab273c-kube-api-access-h74nv\") pod \"cert-manager-webhook-5655c58dd6-rdw9w\" (UID: \"f37f93ad-8593-4bf2-bb4a-7d01e9ab273c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.030413 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8zwg\" (UniqueName: \"kubernetes.io/projected/21c43c6c-8aa0-4558-8e9e-ba34798b6c18-kube-api-access-g8zwg\") pod \"cert-manager-5b446d88c5-w94hv\" (UID: \"21c43c6c-8aa0-4558-8e9e-ba34798b6c18\") " pod="cert-manager/cert-manager-5b446d88c5-w94hv" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.031092 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h74nv\" (UniqueName: \"kubernetes.io/projected/f37f93ad-8593-4bf2-bb4a-7d01e9ab273c-kube-api-access-h74nv\") pod \"cert-manager-webhook-5655c58dd6-rdw9w\" (UID: \"f37f93ad-8593-4bf2-bb4a-7d01e9ab273c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.035075 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcbtw\" (UniqueName: \"kubernetes.io/projected/daf3ed47-3e82-4a2a-ae17-d56aa124499e-kube-api-access-fcbtw\") pod \"cert-manager-cainjector-7f985d654d-8s797\" (UID: \"daf3ed47-3e82-4a2a-ae17-d56aa124499e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-8s797" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.037488 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-8s797" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.045934 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-w94hv" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.066909 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.341003 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rdw9w"] Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.350134 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.486914 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-8s797"] Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.490709 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w94hv"] Oct 04 08:26:50 crc kubenswrapper[4969]: W1004 08:26:50.495507 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf3ed47_3e82_4a2a_ae17_d56aa124499e.slice/crio-10e01e0607bae3dd2cf41e3bcc6ff7825ae8637b02b9f2126423c2419f204c16 WatchSource:0}: Error finding container 10e01e0607bae3dd2cf41e3bcc6ff7825ae8637b02b9f2126423c2419f204c16: Status 404 returned error can't find the container with id 10e01e0607bae3dd2cf41e3bcc6ff7825ae8637b02b9f2126423c2419f204c16 Oct 04 08:26:50 crc kubenswrapper[4969]: W1004 08:26:50.499080 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21c43c6c_8aa0_4558_8e9e_ba34798b6c18.slice/crio-ab536112abdbd853d0068af6fce5f22ddefab2f0d739311b5b190a463ea10105 WatchSource:0}: Error finding container ab536112abdbd853d0068af6fce5f22ddefab2f0d739311b5b190a463ea10105: Status 404 returned error can't find the container with id ab536112abdbd853d0068af6fce5f22ddefab2f0d739311b5b190a463ea10105 Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.921621 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-w94hv" event={"ID":"21c43c6c-8aa0-4558-8e9e-ba34798b6c18","Type":"ContainerStarted","Data":"ab536112abdbd853d0068af6fce5f22ddefab2f0d739311b5b190a463ea10105"} Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.923464 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" event={"ID":"f37f93ad-8593-4bf2-bb4a-7d01e9ab273c","Type":"ContainerStarted","Data":"5f918042a746ab06a604e96edb9401ae414549f81e785ce0aae9a3bb724798fe"} Oct 04 08:26:50 crc kubenswrapper[4969]: I1004 08:26:50.924575 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-8s797" event={"ID":"daf3ed47-3e82-4a2a-ae17-d56aa124499e","Type":"ContainerStarted","Data":"10e01e0607bae3dd2cf41e3bcc6ff7825ae8637b02b9f2126423c2419f204c16"} Oct 04 08:26:52 crc kubenswrapper[4969]: I1004 08:26:52.939773 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" event={"ID":"f37f93ad-8593-4bf2-bb4a-7d01e9ab273c","Type":"ContainerStarted","Data":"69ff81c995582f19c5d6bbe51aeacd6454a0b5e8c40ac610b5b1161ff1e55a2b"} Oct 04 08:26:52 crc kubenswrapper[4969]: I1004 08:26:52.940112 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" Oct 04 08:26:52 crc kubenswrapper[4969]: I1004 08:26:52.952890 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" podStartSLOduration=1.902068818 podStartE2EDuration="3.95286902s" podCreationTimestamp="2025-10-04 08:26:49 +0000 UTC" firstStartedPulling="2025-10-04 08:26:50.349930642 +0000 UTC m=+638.104199456" lastFinishedPulling="2025-10-04 08:26:52.400730794 +0000 UTC m=+640.154999658" observedRunningTime="2025-10-04 08:26:52.951859506 +0000 UTC m=+640.706128340" watchObservedRunningTime="2025-10-04 08:26:52.95286902 +0000 UTC m=+640.707137834" Oct 04 08:26:54 crc kubenswrapper[4969]: I1004 08:26:54.956475 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-8s797" event={"ID":"daf3ed47-3e82-4a2a-ae17-d56aa124499e","Type":"ContainerStarted","Data":"7f07bff4054e0dbfa574b05417514ca2ac9904fc22c3f520d3deaa9c0526c2c4"} Oct 04 08:26:54 crc kubenswrapper[4969]: I1004 08:26:54.959305 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-w94hv" event={"ID":"21c43c6c-8aa0-4558-8e9e-ba34798b6c18","Type":"ContainerStarted","Data":"d06b080e6372e3ef8e6e026772c89fa9a5b33a3acf6ab2a5b4c164698bcb8585"} Oct 04 08:26:54 crc kubenswrapper[4969]: I1004 08:26:54.976538 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-8s797" podStartSLOduration=2.512555668 podStartE2EDuration="5.976514117s" podCreationTimestamp="2025-10-04 08:26:49 +0000 UTC" firstStartedPulling="2025-10-04 08:26:50.49772613 +0000 UTC m=+638.251994944" lastFinishedPulling="2025-10-04 08:26:53.961684569 +0000 UTC m=+641.715953393" observedRunningTime="2025-10-04 08:26:54.971393749 +0000 UTC m=+642.725662583" watchObservedRunningTime="2025-10-04 08:26:54.976514117 +0000 UTC m=+642.730782951" Oct 04 08:26:54 crc kubenswrapper[4969]: I1004 08:26:54.992516 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-w94hv" podStartSLOduration=2.5388481609999998 podStartE2EDuration="5.992489504s" podCreationTimestamp="2025-10-04 08:26:49 +0000 UTC" firstStartedPulling="2025-10-04 08:26:50.50097767 +0000 UTC m=+638.255246484" lastFinishedPulling="2025-10-04 08:26:53.954619013 +0000 UTC m=+641.708887827" observedRunningTime="2025-10-04 08:26:54.990120135 +0000 UTC m=+642.744388999" watchObservedRunningTime="2025-10-04 08:26:54.992489504 +0000 UTC m=+642.746758368" Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.070320 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-rdw9w" Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.298005 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dl8q8"] Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.299963 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovn-controller" containerID="cri-o://193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68" gracePeriod=30 Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.300093 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="northd" containerID="cri-o://e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a" gracePeriod=30 Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.300140 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb" gracePeriod=30 Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.300178 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kube-rbac-proxy-node" containerID="cri-o://d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b" gracePeriod=30 Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.300207 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovn-acl-logging" containerID="cri-o://f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931" gracePeriod=30 Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.299980 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="nbdb" containerID="cri-o://0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40" gracePeriod=30 Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.300745 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="sbdb" containerID="cri-o://1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414" gracePeriod=30 Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.341033 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" containerID="cri-o://a1c38ffbc2e6169e3cdd8e7740ea90186281ace71f537797e1f8c07c93acf1c6" gracePeriod=30 Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.997571 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/4.log" Oct 04 08:27:00 crc kubenswrapper[4969]: I1004 08:27:00.998851 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/3.log" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.002218 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovn-acl-logging/0.log" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.002838 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovn-controller/0.log" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003409 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="a1c38ffbc2e6169e3cdd8e7740ea90186281ace71f537797e1f8c07c93acf1c6" exitCode=2 Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003499 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414" exitCode=0 Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003519 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40" exitCode=0 Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003537 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a" exitCode=0 Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003558 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb" exitCode=0 Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003575 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b" exitCode=0 Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003574 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"a1c38ffbc2e6169e3cdd8e7740ea90186281ace71f537797e1f8c07c93acf1c6"} Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003593 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931" exitCode=143 Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003666 4969 scope.go:117] "RemoveContainer" containerID="bf381d5e0cc1b6d7fabf05a6d4e91eaa94e3098a38bdb84564b7bc3ca852db4d" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003709 4969 generic.go:334] "Generic (PLEG): container finished" podID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerID="193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68" exitCode=143 Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003644 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414"} Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003807 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40"} Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003838 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a"} Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003937 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb"} Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003973 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b"} Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.003991 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931"} Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.004009 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68"} Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.006819 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/2.log" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.007547 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/1.log" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.007610 4969 generic.go:334] "Generic (PLEG): container finished" podID="8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6" containerID="e01ce426ae70129a5c459d559cccad2b6c2be1ece70c9c3f21dfd49482f3cfb1" exitCode=2 Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.007647 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mf8dr" event={"ID":"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6","Type":"ContainerDied","Data":"e01ce426ae70129a5c459d559cccad2b6c2be1ece70c9c3f21dfd49482f3cfb1"} Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.008258 4969 scope.go:117] "RemoveContainer" containerID="e01ce426ae70129a5c459d559cccad2b6c2be1ece70c9c3f21dfd49482f3cfb1" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.008670 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-mf8dr_openshift-multus(8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6)\"" pod="openshift-multus/multus-mf8dr" podUID="8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.133984 4969 scope.go:117] "RemoveContainer" containerID="c038d6119f6cea5038ea95b0398df353d1b360781991bd8905fc325661aea995" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.136345 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/4.log" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.138736 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovn-acl-logging/0.log" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.139298 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovn-controller/0.log" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.139727 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197451 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f78hd"] Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197787 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197804 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197814 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="nbdb" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197821 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="nbdb" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197830 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kube-rbac-proxy-node" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197860 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kube-rbac-proxy-node" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197871 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="sbdb" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197876 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="sbdb" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197889 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197895 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197901 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="northd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197907 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="northd" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197934 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovn-acl-logging" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197941 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovn-acl-logging" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197948 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197955 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197962 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197969 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.197981 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovn-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.197989 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovn-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.198026 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kubecfg-setup" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198035 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kubecfg-setup" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198213 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198230 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="northd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198242 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kube-rbac-proxy-node" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198252 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovn-acl-logging" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198263 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="nbdb" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198298 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198309 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovn-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198320 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="sbdb" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198331 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198339 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.198506 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198518 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: E1004 08:27:01.198526 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198534 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.198721 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.199040 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" containerName="ovnkube-controller" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.201178 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257583 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-ovn-kubernetes\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257629 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257663 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcv2d\" (UniqueName: \"kubernetes.io/projected/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-kube-api-access-bcv2d\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257689 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-env-overrides\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257709 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-var-lib-openvswitch\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257729 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-netns\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257744 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-netd\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257671 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257823 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257838 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257887 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-kubelet\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257925 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257942 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257904 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-slash\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.257981 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-ovn\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258032 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-slash" (OuterVolumeSpecName: "host-slash") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258029 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258097 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258116 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-systemd-units\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258103 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258172 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-script-lib\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258196 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-systemd\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258169 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258211 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-node-log\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258241 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovn-node-metrics-cert\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258257 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-etc-openvswitch\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258276 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-openvswitch\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258293 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-bin\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258326 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-node-log" (OuterVolumeSpecName: "node-log") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258342 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258352 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258365 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258387 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-config\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258403 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-log-socket\") pod \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\" (UID: \"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be\") " Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258504 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-log-socket" (OuterVolumeSpecName: "log-socket") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258640 4969 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258650 4969 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258659 4969 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258667 4969 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-log-socket\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258676 4969 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258685 4969 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258693 4969 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258701 4969 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258709 4969 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258716 4969 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258725 4969 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258732 4969 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-host-slash\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258740 4969 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258747 4969 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258755 4969 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-node-log\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258821 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.258932 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.262946 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-kube-api-access-bcv2d" (OuterVolumeSpecName: "kube-api-access-bcv2d") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "kube-api-access-bcv2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.262987 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.270248 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" (UID: "9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.360364 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-cni-bin\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.360516 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3a5e149a-8f5a-4523-8a47-20f3a61338b3-ovnkube-script-lib\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.360554 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-run-ovn\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.360592 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-run-systemd\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.360627 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-run-openvswitch\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.360678 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-slash\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.360712 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gs8v\" (UniqueName: \"kubernetes.io/projected/3a5e149a-8f5a-4523-8a47-20f3a61338b3-kube-api-access-9gs8v\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.360881 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.360965 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-systemd-units\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361026 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361078 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a5e149a-8f5a-4523-8a47-20f3a61338b3-ovn-node-metrics-cert\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361127 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-etc-openvswitch\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361201 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-cni-netd\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361241 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-var-lib-openvswitch\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361329 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a5e149a-8f5a-4523-8a47-20f3a61338b3-env-overrides\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361633 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-node-log\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361683 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a5e149a-8f5a-4523-8a47-20f3a61338b3-ovnkube-config\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361717 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-log-socket\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361752 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-run-netns\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361805 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-kubelet\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361884 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcv2d\" (UniqueName: \"kubernetes.io/projected/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-kube-api-access-bcv2d\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361919 4969 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361939 4969 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361958 4969 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.361979 4969 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.463564 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-slash\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.463664 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-slash\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.463981 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gs8v\" (UniqueName: \"kubernetes.io/projected/3a5e149a-8f5a-4523-8a47-20f3a61338b3-kube-api-access-9gs8v\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464100 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464147 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-systemd-units\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464184 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464227 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a5e149a-8f5a-4523-8a47-20f3a61338b3-ovn-node-metrics-cert\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464242 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464267 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-etc-openvswitch\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464314 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-etc-openvswitch\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464350 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-cni-netd\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464362 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-systemd-units\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464372 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-var-lib-openvswitch\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464405 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a5e149a-8f5a-4523-8a47-20f3a61338b3-env-overrides\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464404 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464489 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-node-log\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464524 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a5e149a-8f5a-4523-8a47-20f3a61338b3-ovnkube-config\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464548 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-log-socket\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464578 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-run-netns\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464639 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-kubelet\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464713 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-cni-bin\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464734 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3a5e149a-8f5a-4523-8a47-20f3a61338b3-ovnkube-script-lib\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464752 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-run-ovn\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464782 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-run-systemd\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464800 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-run-openvswitch\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464893 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-run-openvswitch\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464931 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-run-netns\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464940 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-log-socket\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464961 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-kubelet\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464993 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-cni-netd\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.464995 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-host-cni-bin\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.465040 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-var-lib-openvswitch\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.465325 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-node-log\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.465691 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a5e149a-8f5a-4523-8a47-20f3a61338b3-env-overrides\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.465729 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-run-ovn\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.465750 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3a5e149a-8f5a-4523-8a47-20f3a61338b3-run-systemd\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.465880 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3a5e149a-8f5a-4523-8a47-20f3a61338b3-ovnkube-script-lib\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.466744 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a5e149a-8f5a-4523-8a47-20f3a61338b3-ovnkube-config\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.469005 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a5e149a-8f5a-4523-8a47-20f3a61338b3-ovn-node-metrics-cert\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.485068 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gs8v\" (UniqueName: \"kubernetes.io/projected/3a5e149a-8f5a-4523-8a47-20f3a61338b3-kube-api-access-9gs8v\") pod \"ovnkube-node-f78hd\" (UID: \"3a5e149a-8f5a-4523-8a47-20f3a61338b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:01 crc kubenswrapper[4969]: I1004 08:27:01.514154 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.016977 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/2.log" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.020649 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovnkube-controller/4.log" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.024036 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovn-acl-logging/0.log" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.024956 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dl8q8_9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/ovn-controller/0.log" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.025578 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" event={"ID":"9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be","Type":"ContainerDied","Data":"91fea041a813870b1ef1f48eca51b8092fe25022c8e2e2d89b8f7463e1519b08"} Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.025644 4969 scope.go:117] "RemoveContainer" containerID="a1c38ffbc2e6169e3cdd8e7740ea90186281ace71f537797e1f8c07c93acf1c6" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.025655 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dl8q8" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.028012 4969 generic.go:334] "Generic (PLEG): container finished" podID="3a5e149a-8f5a-4523-8a47-20f3a61338b3" containerID="3c619a5b500a517dd769895c4b2b083c35d6712e60c225330f9ae346177e247a" exitCode=0 Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.028059 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerDied","Data":"3c619a5b500a517dd769895c4b2b083c35d6712e60c225330f9ae346177e247a"} Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.028160 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerStarted","Data":"6e83e0d403588ef470f47957de151439dec21760a7e67aaab02a9bca2fa3904a"} Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.077796 4969 scope.go:117] "RemoveContainer" containerID="1121e77bf1478a611c82e14d99343f21d99d5c12e392b4aa8b1b0ea15d519414" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.116569 4969 scope.go:117] "RemoveContainer" containerID="0e23d0ffc9ebee08053f9ea269dd906cd065cbec0ee2605b109ea6ab5b6b0b40" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.120283 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dl8q8"] Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.139612 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dl8q8"] Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.140945 4969 scope.go:117] "RemoveContainer" containerID="e5d08ac58048f43fe91b569cb4a3d230c38ff71bd1c8a9388210e803593b6d9a" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.161597 4969 scope.go:117] "RemoveContainer" containerID="01ca77de7b1973ff75beb1e9d8d775a504904fda34d3f83df6f238d4982a5afb" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.180089 4969 scope.go:117] "RemoveContainer" containerID="d6dfad86a90872e6c0d4c5472b1dfc1ae35982694dc77554fb4962b86a4cc93b" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.192114 4969 scope.go:117] "RemoveContainer" containerID="f541fcc367196141abd4e739e827e0f46f1f13d38721c7158b48669512ebb931" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.209100 4969 scope.go:117] "RemoveContainer" containerID="193152cc2939dd9c99f3979cbfe2a52f3071bb89ae44d8870e0173185757dd68" Oct 04 08:27:02 crc kubenswrapper[4969]: I1004 08:27:02.300758 4969 scope.go:117] "RemoveContainer" containerID="a191a5f92e6dfe0db18e81bd16c268cd6e28026fc085f4d1a68e9a78a8ece38d" Oct 04 08:27:03 crc kubenswrapper[4969]: I1004 08:27:03.037873 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerStarted","Data":"95d5217ec8a22b5a09c4caa7a91592369f6daacfdafd941dc16229068cdbcb1e"} Oct 04 08:27:03 crc kubenswrapper[4969]: I1004 08:27:03.038157 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerStarted","Data":"6c104e31aab7e66beb231ff2469511114f51c1fb8d3c9a788ea1432797928440"} Oct 04 08:27:03 crc kubenswrapper[4969]: I1004 08:27:03.038167 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerStarted","Data":"a03e292db8b475ae1d39949041219dc52f0eab138c3b483d915140b24a0e57fb"} Oct 04 08:27:03 crc kubenswrapper[4969]: I1004 08:27:03.038176 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerStarted","Data":"6da78cc582ecbdcabd2e948176135f9b3de0d16cd559d8520c4e2a29f26e0169"} Oct 04 08:27:03 crc kubenswrapper[4969]: I1004 08:27:03.038184 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerStarted","Data":"acc52386f36be023ef186e38da48c54dffb5c53c4ffa9a502684256ebaa56b60"} Oct 04 08:27:03 crc kubenswrapper[4969]: I1004 08:27:03.038196 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerStarted","Data":"e686f7c81c99d3b7a372b0fac0c1f542ad1abaf85b1f3d51e504114ac1f98f5f"} Oct 04 08:27:03 crc kubenswrapper[4969]: I1004 08:27:03.060283 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be" path="/var/lib/kubelet/pods/9e5258c5-cab0-4d0b-b5fd-0e8ffd1f19be/volumes" Oct 04 08:27:06 crc kubenswrapper[4969]: I1004 08:27:06.064490 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerStarted","Data":"4a1e46b27fd1eb547e766daf2a3ba77c52b6c24fea8505259420f2f84c8970bf"} Oct 04 08:27:08 crc kubenswrapper[4969]: I1004 08:27:08.080532 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" event={"ID":"3a5e149a-8f5a-4523-8a47-20f3a61338b3","Type":"ContainerStarted","Data":"8b5c2222c52076b51e4d50ef070c63cc7a5a3c430be93f989046ce72195fc943"} Oct 04 08:27:08 crc kubenswrapper[4969]: I1004 08:27:08.124664 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" podStartSLOduration=7.124648335 podStartE2EDuration="7.124648335s" podCreationTimestamp="2025-10-04 08:27:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:27:08.121759484 +0000 UTC m=+655.876028338" watchObservedRunningTime="2025-10-04 08:27:08.124648335 +0000 UTC m=+655.878917139" Oct 04 08:27:09 crc kubenswrapper[4969]: I1004 08:27:09.088204 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:09 crc kubenswrapper[4969]: I1004 08:27:09.088261 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:09 crc kubenswrapper[4969]: I1004 08:27:09.088276 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:09 crc kubenswrapper[4969]: I1004 08:27:09.169232 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:09 crc kubenswrapper[4969]: I1004 08:27:09.170979 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:16 crc kubenswrapper[4969]: I1004 08:27:16.054953 4969 scope.go:117] "RemoveContainer" containerID="e01ce426ae70129a5c459d559cccad2b6c2be1ece70c9c3f21dfd49482f3cfb1" Oct 04 08:27:16 crc kubenswrapper[4969]: E1004 08:27:16.055514 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-mf8dr_openshift-multus(8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6)\"" pod="openshift-multus/multus-mf8dr" podUID="8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6" Oct 04 08:27:28 crc kubenswrapper[4969]: I1004 08:27:28.055749 4969 scope.go:117] "RemoveContainer" containerID="e01ce426ae70129a5c459d559cccad2b6c2be1ece70c9c3f21dfd49482f3cfb1" Oct 04 08:27:28 crc kubenswrapper[4969]: I1004 08:27:28.205962 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/2.log" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.215823 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mf8dr_8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6/kube-multus/2.log" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.215908 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mf8dr" event={"ID":"8f4f1158-78b1-4759-b3b1-a0cfd8ee1ba6","Type":"ContainerStarted","Data":"f1b7eca32f94a206270a3a8917a28e200f0660c95ae2f51fda1d353236e3191b"} Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.276686 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw"] Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.278473 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.281500 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.294056 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw"] Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.457572 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7m9d\" (UniqueName: \"kubernetes.io/projected/82b1344f-981d-4caf-82c2-c8035d9fb68b-kube-api-access-q7m9d\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.457780 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.457848 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.559543 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.559610 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7m9d\" (UniqueName: \"kubernetes.io/projected/82b1344f-981d-4caf-82c2-c8035d9fb68b-kube-api-access-q7m9d\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.559702 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.560292 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.560333 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.604276 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7m9d\" (UniqueName: \"kubernetes.io/projected/82b1344f-981d-4caf-82c2-c8035d9fb68b-kube-api-access-q7m9d\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: I1004 08:27:29.615491 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: E1004 08:27:29.659332 4969 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_openshift-marketplace_82b1344f-981d-4caf-82c2-c8035d9fb68b_0(966d739b5010845a9dbd95fb1717f0f634a877c79181302ed720664998404752): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 08:27:29 crc kubenswrapper[4969]: E1004 08:27:29.659514 4969 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_openshift-marketplace_82b1344f-981d-4caf-82c2-c8035d9fb68b_0(966d739b5010845a9dbd95fb1717f0f634a877c79181302ed720664998404752): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: E1004 08:27:29.659573 4969 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_openshift-marketplace_82b1344f-981d-4caf-82c2-c8035d9fb68b_0(966d739b5010845a9dbd95fb1717f0f634a877c79181302ed720664998404752): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:29 crc kubenswrapper[4969]: E1004 08:27:29.659663 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_openshift-marketplace(82b1344f-981d-4caf-82c2-c8035d9fb68b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_openshift-marketplace(82b1344f-981d-4caf-82c2-c8035d9fb68b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_openshift-marketplace_82b1344f-981d-4caf-82c2-c8035d9fb68b_0(966d739b5010845a9dbd95fb1717f0f634a877c79181302ed720664998404752): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" podUID="82b1344f-981d-4caf-82c2-c8035d9fb68b" Oct 04 08:27:30 crc kubenswrapper[4969]: I1004 08:27:30.222592 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:30 crc kubenswrapper[4969]: I1004 08:27:30.223590 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:30 crc kubenswrapper[4969]: I1004 08:27:30.460071 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw"] Oct 04 08:27:30 crc kubenswrapper[4969]: W1004 08:27:30.466032 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82b1344f_981d_4caf_82c2_c8035d9fb68b.slice/crio-0bc433cf4ec64b3ea13a8ed43d3325b3795e65fc1d787d982379b60c1f557b2d WatchSource:0}: Error finding container 0bc433cf4ec64b3ea13a8ed43d3325b3795e65fc1d787d982379b60c1f557b2d: Status 404 returned error can't find the container with id 0bc433cf4ec64b3ea13a8ed43d3325b3795e65fc1d787d982379b60c1f557b2d Oct 04 08:27:31 crc kubenswrapper[4969]: I1004 08:27:31.230696 4969 generic.go:334] "Generic (PLEG): container finished" podID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerID="0ba618d3aef5b90c32b954baa1f1a5bafbfcbf4cfcddce1a4c2bb0894705b625" exitCode=0 Oct 04 08:27:31 crc kubenswrapper[4969]: I1004 08:27:31.230737 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" event={"ID":"82b1344f-981d-4caf-82c2-c8035d9fb68b","Type":"ContainerDied","Data":"0ba618d3aef5b90c32b954baa1f1a5bafbfcbf4cfcddce1a4c2bb0894705b625"} Oct 04 08:27:31 crc kubenswrapper[4969]: I1004 08:27:31.230762 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" event={"ID":"82b1344f-981d-4caf-82c2-c8035d9fb68b","Type":"ContainerStarted","Data":"0bc433cf4ec64b3ea13a8ed43d3325b3795e65fc1d787d982379b60c1f557b2d"} Oct 04 08:27:31 crc kubenswrapper[4969]: I1004 08:27:31.535749 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f78hd" Oct 04 08:27:33 crc kubenswrapper[4969]: I1004 08:27:33.243280 4969 generic.go:334] "Generic (PLEG): container finished" podID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerID="d305c6363e82729dce5bec27f2767ac20ef6f803b81047b5845a6f47e28b7c20" exitCode=0 Oct 04 08:27:33 crc kubenswrapper[4969]: I1004 08:27:33.243372 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" event={"ID":"82b1344f-981d-4caf-82c2-c8035d9fb68b","Type":"ContainerDied","Data":"d305c6363e82729dce5bec27f2767ac20ef6f803b81047b5845a6f47e28b7c20"} Oct 04 08:27:34 crc kubenswrapper[4969]: I1004 08:27:34.266867 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" event={"ID":"82b1344f-981d-4caf-82c2-c8035d9fb68b","Type":"ContainerStarted","Data":"f9eb1e5b80b74bd02a8b316d5843f51c8e8cc72a6f7d0283da93d54352373f2f"} Oct 04 08:27:34 crc kubenswrapper[4969]: I1004 08:27:34.297524 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" podStartSLOduration=4.209146097 podStartE2EDuration="5.297498134s" podCreationTimestamp="2025-10-04 08:27:29 +0000 UTC" firstStartedPulling="2025-10-04 08:27:31.232125673 +0000 UTC m=+678.986394487" lastFinishedPulling="2025-10-04 08:27:32.32047772 +0000 UTC m=+680.074746524" observedRunningTime="2025-10-04 08:27:34.291993558 +0000 UTC m=+682.046262432" watchObservedRunningTime="2025-10-04 08:27:34.297498134 +0000 UTC m=+682.051766988" Oct 04 08:27:35 crc kubenswrapper[4969]: I1004 08:27:35.276338 4969 generic.go:334] "Generic (PLEG): container finished" podID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerID="f9eb1e5b80b74bd02a8b316d5843f51c8e8cc72a6f7d0283da93d54352373f2f" exitCode=0 Oct 04 08:27:35 crc kubenswrapper[4969]: I1004 08:27:35.276386 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" event={"ID":"82b1344f-981d-4caf-82c2-c8035d9fb68b","Type":"ContainerDied","Data":"f9eb1e5b80b74bd02a8b316d5843f51c8e8cc72a6f7d0283da93d54352373f2f"} Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.612691 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.752117 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7m9d\" (UniqueName: \"kubernetes.io/projected/82b1344f-981d-4caf-82c2-c8035d9fb68b-kube-api-access-q7m9d\") pod \"82b1344f-981d-4caf-82c2-c8035d9fb68b\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.752259 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-util\") pod \"82b1344f-981d-4caf-82c2-c8035d9fb68b\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.752335 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-bundle\") pod \"82b1344f-981d-4caf-82c2-c8035d9fb68b\" (UID: \"82b1344f-981d-4caf-82c2-c8035d9fb68b\") " Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.755828 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-bundle" (OuterVolumeSpecName: "bundle") pod "82b1344f-981d-4caf-82c2-c8035d9fb68b" (UID: "82b1344f-981d-4caf-82c2-c8035d9fb68b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.760761 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82b1344f-981d-4caf-82c2-c8035d9fb68b-kube-api-access-q7m9d" (OuterVolumeSpecName: "kube-api-access-q7m9d") pod "82b1344f-981d-4caf-82c2-c8035d9fb68b" (UID: "82b1344f-981d-4caf-82c2-c8035d9fb68b"). InnerVolumeSpecName "kube-api-access-q7m9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.778160 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-util" (OuterVolumeSpecName: "util") pod "82b1344f-981d-4caf-82c2-c8035d9fb68b" (UID: "82b1344f-981d-4caf-82c2-c8035d9fb68b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.857403 4969 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-util\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.857494 4969 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b1344f-981d-4caf-82c2-c8035d9fb68b-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:36 crc kubenswrapper[4969]: I1004 08:27:36.857521 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7m9d\" (UniqueName: \"kubernetes.io/projected/82b1344f-981d-4caf-82c2-c8035d9fb68b-kube-api-access-q7m9d\") on node \"crc\" DevicePath \"\"" Oct 04 08:27:37 crc kubenswrapper[4969]: I1004 08:27:37.292924 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" event={"ID":"82b1344f-981d-4caf-82c2-c8035d9fb68b","Type":"ContainerDied","Data":"0bc433cf4ec64b3ea13a8ed43d3325b3795e65fc1d787d982379b60c1f557b2d"} Oct 04 08:27:37 crc kubenswrapper[4969]: I1004 08:27:37.293335 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bc433cf4ec64b3ea13a8ed43d3325b3795e65fc1d787d982379b60c1f557b2d" Oct 04 08:27:37 crc kubenswrapper[4969]: I1004 08:27:37.293024 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.319478 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb"] Oct 04 08:27:46 crc kubenswrapper[4969]: E1004 08:27:46.321926 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerName="util" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.322038 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerName="util" Oct 04 08:27:46 crc kubenswrapper[4969]: E1004 08:27:46.322128 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerName="extract" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.322226 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerName="extract" Oct 04 08:27:46 crc kubenswrapper[4969]: E1004 08:27:46.322331 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerName="pull" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.322449 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerName="pull" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.322743 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="82b1344f-981d-4caf-82c2-c8035d9fb68b" containerName="extract" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.323541 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.326144 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.326330 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.327163 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-2tv4n" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.358457 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.380593 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4jqr\" (UniqueName: \"kubernetes.io/projected/91044ecd-9c81-4052-9e16-9224f04f079e-kube-api-access-l4jqr\") pod \"obo-prometheus-operator-7c8cf85677-2xmcb\" (UID: \"91044ecd-9c81-4052-9e16-9224f04f079e\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.445384 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.446569 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.448646 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.448917 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-llc9w" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.450321 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.451083 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.463634 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.481172 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn\" (UID: \"c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.481258 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4jqr\" (UniqueName: \"kubernetes.io/projected/91044ecd-9c81-4052-9e16-9224f04f079e-kube-api-access-l4jqr\") pod \"obo-prometheus-operator-7c8cf85677-2xmcb\" (UID: \"91044ecd-9c81-4052-9e16-9224f04f079e\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.481307 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fca6110d-f0fb-4b9e-8c48-749c188c96a3-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6\" (UID: \"fca6110d-f0fb-4b9e-8c48-749c188c96a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.481364 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fca6110d-f0fb-4b9e-8c48-749c188c96a3-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6\" (UID: \"fca6110d-f0fb-4b9e-8c48-749c188c96a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.481389 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn\" (UID: \"c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.486263 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.505732 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4jqr\" (UniqueName: \"kubernetes.io/projected/91044ecd-9c81-4052-9e16-9224f04f079e-kube-api-access-l4jqr\") pod \"obo-prometheus-operator-7c8cf85677-2xmcb\" (UID: \"91044ecd-9c81-4052-9e16-9224f04f079e\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.560067 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-vl7q6"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.560998 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.562915 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-njd4w" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.563103 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.582763 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fca6110d-f0fb-4b9e-8c48-749c188c96a3-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6\" (UID: \"fca6110d-f0fb-4b9e-8c48-749c188c96a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.582846 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkxrs\" (UniqueName: \"kubernetes.io/projected/15b3f66e-d095-42c3-94e2-b9ee1071d638-kube-api-access-pkxrs\") pod \"observability-operator-cc5f78dfc-vl7q6\" (UID: \"15b3f66e-d095-42c3-94e2-b9ee1071d638\") " pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.582877 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fca6110d-f0fb-4b9e-8c48-749c188c96a3-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6\" (UID: \"fca6110d-f0fb-4b9e-8c48-749c188c96a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.582901 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn\" (UID: \"c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.582929 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn\" (UID: \"c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.582953 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/15b3f66e-d095-42c3-94e2-b9ee1071d638-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-vl7q6\" (UID: \"15b3f66e-d095-42c3-94e2-b9ee1071d638\") " pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.586873 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fca6110d-f0fb-4b9e-8c48-749c188c96a3-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6\" (UID: \"fca6110d-f0fb-4b9e-8c48-749c188c96a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.587635 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fca6110d-f0fb-4b9e-8c48-749c188c96a3-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6\" (UID: \"fca6110d-f0fb-4b9e-8c48-749c188c96a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.590472 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-vl7q6"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.593986 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn\" (UID: \"c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.597467 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn\" (UID: \"c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.644610 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-2c6cm"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.645377 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.648744 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-brqgc" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.653680 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.655192 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-2c6cm"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.684322 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/15b3f66e-d095-42c3-94e2-b9ee1071d638-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-vl7q6\" (UID: \"15b3f66e-d095-42c3-94e2-b9ee1071d638\") " pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.684905 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpg5d\" (UniqueName: \"kubernetes.io/projected/46090d90-0cf0-47f4-a19f-ad5aa65df521-kube-api-access-vpg5d\") pod \"perses-operator-54bc95c9fb-2c6cm\" (UID: \"46090d90-0cf0-47f4-a19f-ad5aa65df521\") " pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.684952 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/46090d90-0cf0-47f4-a19f-ad5aa65df521-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-2c6cm\" (UID: \"46090d90-0cf0-47f4-a19f-ad5aa65df521\") " pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.685003 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkxrs\" (UniqueName: \"kubernetes.io/projected/15b3f66e-d095-42c3-94e2-b9ee1071d638-kube-api-access-pkxrs\") pod \"observability-operator-cc5f78dfc-vl7q6\" (UID: \"15b3f66e-d095-42c3-94e2-b9ee1071d638\") " pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.694118 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/15b3f66e-d095-42c3-94e2-b9ee1071d638-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-vl7q6\" (UID: \"15b3f66e-d095-42c3-94e2-b9ee1071d638\") " pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.704991 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkxrs\" (UniqueName: \"kubernetes.io/projected/15b3f66e-d095-42c3-94e2-b9ee1071d638-kube-api-access-pkxrs\") pod \"observability-operator-cc5f78dfc-vl7q6\" (UID: \"15b3f66e-d095-42c3-94e2-b9ee1071d638\") " pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.764729 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.770793 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.786065 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/46090d90-0cf0-47f4-a19f-ad5aa65df521-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-2c6cm\" (UID: \"46090d90-0cf0-47f4-a19f-ad5aa65df521\") " pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.786500 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpg5d\" (UniqueName: \"kubernetes.io/projected/46090d90-0cf0-47f4-a19f-ad5aa65df521-kube-api-access-vpg5d\") pod \"perses-operator-54bc95c9fb-2c6cm\" (UID: \"46090d90-0cf0-47f4-a19f-ad5aa65df521\") " pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.787695 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/46090d90-0cf0-47f4-a19f-ad5aa65df521-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-2c6cm\" (UID: \"46090d90-0cf0-47f4-a19f-ad5aa65df521\") " pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.816281 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpg5d\" (UniqueName: \"kubernetes.io/projected/46090d90-0cf0-47f4-a19f-ad5aa65df521-kube-api-access-vpg5d\") pod \"perses-operator-54bc95c9fb-2c6cm\" (UID: \"46090d90-0cf0-47f4-a19f-ad5aa65df521\") " pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.875845 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.890160 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb"] Oct 04 08:27:46 crc kubenswrapper[4969]: I1004 08:27:46.969598 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:27:47 crc kubenswrapper[4969]: I1004 08:27:47.018823 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn"] Oct 04 08:27:47 crc kubenswrapper[4969]: W1004 08:27:47.027746 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7ba6dc9_cfcc_4e19_8502_b5ef4c981b1e.slice/crio-f467d7a320bd7f32b67ce384e547c8e2a70fc01d5c7b15b12d946c454ac9aa35 WatchSource:0}: Error finding container f467d7a320bd7f32b67ce384e547c8e2a70fc01d5c7b15b12d946c454ac9aa35: Status 404 returned error can't find the container with id f467d7a320bd7f32b67ce384e547c8e2a70fc01d5c7b15b12d946c454ac9aa35 Oct 04 08:27:47 crc kubenswrapper[4969]: I1004 08:27:47.153044 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-vl7q6"] Oct 04 08:27:47 crc kubenswrapper[4969]: W1004 08:27:47.159198 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15b3f66e_d095_42c3_94e2_b9ee1071d638.slice/crio-7ac3dccf93fdf37ffef7a6278cbb8b5027c3201efe9b06a93b08c6b2059b0e8e WatchSource:0}: Error finding container 7ac3dccf93fdf37ffef7a6278cbb8b5027c3201efe9b06a93b08c6b2059b0e8e: Status 404 returned error can't find the container with id 7ac3dccf93fdf37ffef7a6278cbb8b5027c3201efe9b06a93b08c6b2059b0e8e Oct 04 08:27:47 crc kubenswrapper[4969]: I1004 08:27:47.231013 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-2c6cm"] Oct 04 08:27:47 crc kubenswrapper[4969]: W1004 08:27:47.237574 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46090d90_0cf0_47f4_a19f_ad5aa65df521.slice/crio-0e28115951a96e4c954a7745ee21d110ac120ff7e3215ab67d91f369a82e98ff WatchSource:0}: Error finding container 0e28115951a96e4c954a7745ee21d110ac120ff7e3215ab67d91f369a82e98ff: Status 404 returned error can't find the container with id 0e28115951a96e4c954a7745ee21d110ac120ff7e3215ab67d91f369a82e98ff Oct 04 08:27:47 crc kubenswrapper[4969]: I1004 08:27:47.261821 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6"] Oct 04 08:27:47 crc kubenswrapper[4969]: W1004 08:27:47.268612 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfca6110d_f0fb_4b9e_8c48_749c188c96a3.slice/crio-9fa1e3e4374d6afbe41762746fdbc41a3bf433b0971a1239d692a79c112caf21 WatchSource:0}: Error finding container 9fa1e3e4374d6afbe41762746fdbc41a3bf433b0971a1239d692a79c112caf21: Status 404 returned error can't find the container with id 9fa1e3e4374d6afbe41762746fdbc41a3bf433b0971a1239d692a79c112caf21 Oct 04 08:27:47 crc kubenswrapper[4969]: I1004 08:27:47.366439 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" event={"ID":"46090d90-0cf0-47f4-a19f-ad5aa65df521","Type":"ContainerStarted","Data":"0e28115951a96e4c954a7745ee21d110ac120ff7e3215ab67d91f369a82e98ff"} Oct 04 08:27:47 crc kubenswrapper[4969]: I1004 08:27:47.367577 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" event={"ID":"15b3f66e-d095-42c3-94e2-b9ee1071d638","Type":"ContainerStarted","Data":"7ac3dccf93fdf37ffef7a6278cbb8b5027c3201efe9b06a93b08c6b2059b0e8e"} Oct 04 08:27:47 crc kubenswrapper[4969]: I1004 08:27:47.368698 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" event={"ID":"fca6110d-f0fb-4b9e-8c48-749c188c96a3","Type":"ContainerStarted","Data":"9fa1e3e4374d6afbe41762746fdbc41a3bf433b0971a1239d692a79c112caf21"} Oct 04 08:27:47 crc kubenswrapper[4969]: I1004 08:27:47.369743 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" event={"ID":"c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e","Type":"ContainerStarted","Data":"f467d7a320bd7f32b67ce384e547c8e2a70fc01d5c7b15b12d946c454ac9aa35"} Oct 04 08:27:47 crc kubenswrapper[4969]: I1004 08:27:47.370931 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb" event={"ID":"91044ecd-9c81-4052-9e16-9224f04f079e","Type":"ContainerStarted","Data":"b2d7c9c065bb2e72676013ecc3608d26ba0b09f992606a48fed7e46e4e9efee7"} Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.479133 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" event={"ID":"46090d90-0cf0-47f4-a19f-ad5aa65df521","Type":"ContainerStarted","Data":"b4692468f6a6be722253450a475b5838c4f3fadd5746180635a2e53cf9e199b3"} Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.480067 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.485895 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" event={"ID":"15b3f66e-d095-42c3-94e2-b9ee1071d638","Type":"ContainerStarted","Data":"8e2b8777a3655e9735f492f7d034c99a45d68975ae992d91049c1eb8d11b503e"} Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.486088 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.490696 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" event={"ID":"fca6110d-f0fb-4b9e-8c48-749c188c96a3","Type":"ContainerStarted","Data":"fa1adc680ddd5ff7e8cfcf3be16d480ada6441f1f7dfa38186c8fb23eb8157d3"} Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.492648 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" event={"ID":"c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e","Type":"ContainerStarted","Data":"18fb7eac9df246899b21b523067aed7df31b2fe7deec275869e5ea338112f5b0"} Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.498499 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" podStartSLOduration=1.565081115 podStartE2EDuration="15.498483378s" podCreationTimestamp="2025-10-04 08:27:46 +0000 UTC" firstStartedPulling="2025-10-04 08:27:47.241320539 +0000 UTC m=+694.995589343" lastFinishedPulling="2025-10-04 08:28:01.174722792 +0000 UTC m=+708.928991606" observedRunningTime="2025-10-04 08:28:01.497227127 +0000 UTC m=+709.251495951" watchObservedRunningTime="2025-10-04 08:28:01.498483378 +0000 UTC m=+709.252752192" Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.515887 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" podStartSLOduration=1.506406504 podStartE2EDuration="15.515875117s" podCreationTimestamp="2025-10-04 08:27:46 +0000 UTC" firstStartedPulling="2025-10-04 08:27:47.16607485 +0000 UTC m=+694.920343664" lastFinishedPulling="2025-10-04 08:28:01.175543463 +0000 UTC m=+708.929812277" observedRunningTime="2025-10-04 08:28:01.51315632 +0000 UTC m=+709.267425134" watchObservedRunningTime="2025-10-04 08:28:01.515875117 +0000 UTC m=+709.270143931" Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.530162 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn" podStartSLOduration=1.435743801 podStartE2EDuration="15.530141199s" podCreationTimestamp="2025-10-04 08:27:46 +0000 UTC" firstStartedPulling="2025-10-04 08:27:47.033717713 +0000 UTC m=+694.787986527" lastFinishedPulling="2025-10-04 08:28:01.128115111 +0000 UTC m=+708.882383925" observedRunningTime="2025-10-04 08:28:01.527812392 +0000 UTC m=+709.282081216" watchObservedRunningTime="2025-10-04 08:28:01.530141199 +0000 UTC m=+709.284410023" Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.567245 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6" podStartSLOduration=1.67536883 podStartE2EDuration="15.567225386s" podCreationTimestamp="2025-10-04 08:27:46 +0000 UTC" firstStartedPulling="2025-10-04 08:27:47.271776942 +0000 UTC m=+695.026045756" lastFinishedPulling="2025-10-04 08:28:01.163633498 +0000 UTC m=+708.917902312" observedRunningTime="2025-10-04 08:28:01.563008091 +0000 UTC m=+709.317276925" watchObservedRunningTime="2025-10-04 08:28:01.567225386 +0000 UTC m=+709.321494200" Oct 04 08:28:01 crc kubenswrapper[4969]: I1004 08:28:01.593796 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-vl7q6" Oct 04 08:28:02 crc kubenswrapper[4969]: I1004 08:28:02.501213 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb" event={"ID":"91044ecd-9c81-4052-9e16-9224f04f079e","Type":"ContainerStarted","Data":"6b2e1d1ad3c5fc07800c53266cba778bc2182610b3e45fc3e677577b3265aa76"} Oct 04 08:28:06 crc kubenswrapper[4969]: I1004 08:28:06.972078 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-2c6cm" Oct 04 08:28:06 crc kubenswrapper[4969]: I1004 08:28:06.996711 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-2xmcb" podStartSLOduration=6.749703721 podStartE2EDuration="20.996678117s" podCreationTimestamp="2025-10-04 08:27:46 +0000 UTC" firstStartedPulling="2025-10-04 08:27:46.91658541 +0000 UTC m=+694.670854224" lastFinishedPulling="2025-10-04 08:28:01.163559806 +0000 UTC m=+708.917828620" observedRunningTime="2025-10-04 08:28:02.53802204 +0000 UTC m=+710.292290884" watchObservedRunningTime="2025-10-04 08:28:06.996678117 +0000 UTC m=+714.750946981" Oct 04 08:28:19 crc kubenswrapper[4969]: I1004 08:28:19.666847 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:28:19 crc kubenswrapper[4969]: I1004 08:28:19.669319 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.568161 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp"] Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.569346 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.572489 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.599915 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp"] Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.697760 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.697825 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.697871 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hxbg\" (UniqueName: \"kubernetes.io/projected/12c98149-2fa0-4b54-9940-92d1a818024c-kube-api-access-5hxbg\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.799249 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.799313 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.799376 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hxbg\" (UniqueName: \"kubernetes.io/projected/12c98149-2fa0-4b54-9940-92d1a818024c-kube-api-access-5hxbg\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.799978 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.800394 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.823500 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hxbg\" (UniqueName: \"kubernetes.io/projected/12c98149-2fa0-4b54-9940-92d1a818024c-kube-api-access-5hxbg\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:25 crc kubenswrapper[4969]: I1004 08:28:25.890718 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:26 crc kubenswrapper[4969]: I1004 08:28:26.111508 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp"] Oct 04 08:28:26 crc kubenswrapper[4969]: I1004 08:28:26.688283 4969 generic.go:334] "Generic (PLEG): container finished" podID="12c98149-2fa0-4b54-9940-92d1a818024c" containerID="c2f97a4ae6436cffddc015818009dd7278abd1c707fa6b269daceef30a1cb07b" exitCode=0 Oct 04 08:28:26 crc kubenswrapper[4969]: I1004 08:28:26.688343 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" event={"ID":"12c98149-2fa0-4b54-9940-92d1a818024c","Type":"ContainerDied","Data":"c2f97a4ae6436cffddc015818009dd7278abd1c707fa6b269daceef30a1cb07b"} Oct 04 08:28:26 crc kubenswrapper[4969]: I1004 08:28:26.688686 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" event={"ID":"12c98149-2fa0-4b54-9940-92d1a818024c","Type":"ContainerStarted","Data":"b2105a342c982bb4e954ec6e61118a3af0ffde06bdddb579feaf7777eeb5e91a"} Oct 04 08:28:28 crc kubenswrapper[4969]: I1004 08:28:28.701835 4969 generic.go:334] "Generic (PLEG): container finished" podID="12c98149-2fa0-4b54-9940-92d1a818024c" containerID="6a3e7ea25efecbdcf57019eed9313a37d05edef7a1d533c4937ec0c06319c472" exitCode=0 Oct 04 08:28:28 crc kubenswrapper[4969]: I1004 08:28:28.701907 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" event={"ID":"12c98149-2fa0-4b54-9940-92d1a818024c","Type":"ContainerDied","Data":"6a3e7ea25efecbdcf57019eed9313a37d05edef7a1d533c4937ec0c06319c472"} Oct 04 08:28:31 crc kubenswrapper[4969]: I1004 08:28:31.729950 4969 generic.go:334] "Generic (PLEG): container finished" podID="12c98149-2fa0-4b54-9940-92d1a818024c" containerID="b13bfdc415e36a026f2a1af36c25aa75d4da3096cb4433110dd9ddc66c26ae2c" exitCode=0 Oct 04 08:28:31 crc kubenswrapper[4969]: I1004 08:28:31.730308 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" event={"ID":"12c98149-2fa0-4b54-9940-92d1a818024c","Type":"ContainerDied","Data":"b13bfdc415e36a026f2a1af36c25aa75d4da3096cb4433110dd9ddc66c26ae2c"} Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.021178 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.105675 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-util\") pod \"12c98149-2fa0-4b54-9940-92d1a818024c\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.105741 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hxbg\" (UniqueName: \"kubernetes.io/projected/12c98149-2fa0-4b54-9940-92d1a818024c-kube-api-access-5hxbg\") pod \"12c98149-2fa0-4b54-9940-92d1a818024c\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.105765 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-bundle\") pod \"12c98149-2fa0-4b54-9940-92d1a818024c\" (UID: \"12c98149-2fa0-4b54-9940-92d1a818024c\") " Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.106959 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-bundle" (OuterVolumeSpecName: "bundle") pod "12c98149-2fa0-4b54-9940-92d1a818024c" (UID: "12c98149-2fa0-4b54-9940-92d1a818024c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.112606 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c98149-2fa0-4b54-9940-92d1a818024c-kube-api-access-5hxbg" (OuterVolumeSpecName: "kube-api-access-5hxbg") pod "12c98149-2fa0-4b54-9940-92d1a818024c" (UID: "12c98149-2fa0-4b54-9940-92d1a818024c"). InnerVolumeSpecName "kube-api-access-5hxbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.116750 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-util" (OuterVolumeSpecName: "util") pod "12c98149-2fa0-4b54-9940-92d1a818024c" (UID: "12c98149-2fa0-4b54-9940-92d1a818024c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.206757 4969 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-util\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.206801 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hxbg\" (UniqueName: \"kubernetes.io/projected/12c98149-2fa0-4b54-9940-92d1a818024c-kube-api-access-5hxbg\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.206816 4969 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12c98149-2fa0-4b54-9940-92d1a818024c-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.752567 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" event={"ID":"12c98149-2fa0-4b54-9940-92d1a818024c","Type":"ContainerDied","Data":"b2105a342c982bb4e954ec6e61118a3af0ffde06bdddb579feaf7777eeb5e91a"} Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.753071 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2105a342c982bb4e954ec6e61118a3af0ffde06bdddb579feaf7777eeb5e91a" Oct 04 08:28:33 crc kubenswrapper[4969]: I1004 08:28:33.752668 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.100105 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf"] Oct 04 08:28:37 crc kubenswrapper[4969]: E1004 08:28:37.100700 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c98149-2fa0-4b54-9940-92d1a818024c" containerName="extract" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.100722 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c98149-2fa0-4b54-9940-92d1a818024c" containerName="extract" Oct 04 08:28:37 crc kubenswrapper[4969]: E1004 08:28:37.100746 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c98149-2fa0-4b54-9940-92d1a818024c" containerName="util" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.100756 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c98149-2fa0-4b54-9940-92d1a818024c" containerName="util" Oct 04 08:28:37 crc kubenswrapper[4969]: E1004 08:28:37.100926 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c98149-2fa0-4b54-9940-92d1a818024c" containerName="pull" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.100938 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c98149-2fa0-4b54-9940-92d1a818024c" containerName="pull" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.101096 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c98149-2fa0-4b54-9940-92d1a818024c" containerName="extract" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.101662 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.104788 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-6qdwj" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.104964 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.104966 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.112643 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf"] Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.266739 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc8nn\" (UniqueName: \"kubernetes.io/projected/005b44cf-b29b-4c9d-aabf-684a15eaba71-kube-api-access-dc8nn\") pod \"nmstate-operator-858ddd8f98-p5kmf\" (UID: \"005b44cf-b29b-4c9d-aabf-684a15eaba71\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.368321 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc8nn\" (UniqueName: \"kubernetes.io/projected/005b44cf-b29b-4c9d-aabf-684a15eaba71-kube-api-access-dc8nn\") pod \"nmstate-operator-858ddd8f98-p5kmf\" (UID: \"005b44cf-b29b-4c9d-aabf-684a15eaba71\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.389374 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc8nn\" (UniqueName: \"kubernetes.io/projected/005b44cf-b29b-4c9d-aabf-684a15eaba71-kube-api-access-dc8nn\") pod \"nmstate-operator-858ddd8f98-p5kmf\" (UID: \"005b44cf-b29b-4c9d-aabf-684a15eaba71\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.416390 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf" Oct 04 08:28:37 crc kubenswrapper[4969]: I1004 08:28:37.878063 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf"] Oct 04 08:28:38 crc kubenswrapper[4969]: I1004 08:28:38.787226 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf" event={"ID":"005b44cf-b29b-4c9d-aabf-684a15eaba71","Type":"ContainerStarted","Data":"2c84e4e01f2eac63c421774b02426c03ddea86045b33be485e371cf4bcd13bfb"} Oct 04 08:28:40 crc kubenswrapper[4969]: I1004 08:28:40.800256 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf" event={"ID":"005b44cf-b29b-4c9d-aabf-684a15eaba71","Type":"ContainerStarted","Data":"add0d481787b4c39d2ff8d2f7aa837f59d772b6c368dc067f72254977c791d27"} Oct 04 08:28:40 crc kubenswrapper[4969]: I1004 08:28:40.824115 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p5kmf" podStartSLOduration=1.462309587 podStartE2EDuration="3.824086071s" podCreationTimestamp="2025-10-04 08:28:37 +0000 UTC" firstStartedPulling="2025-10-04 08:28:37.883055067 +0000 UTC m=+745.637323871" lastFinishedPulling="2025-10-04 08:28:40.244831541 +0000 UTC m=+747.999100355" observedRunningTime="2025-10-04 08:28:40.81800753 +0000 UTC m=+748.572276344" watchObservedRunningTime="2025-10-04 08:28:40.824086071 +0000 UTC m=+748.578354925" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.245021 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lrtsv"] Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.245721 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" podUID="846186fd-c540-464d-b85a-a67f0c8870b6" containerName="controller-manager" containerID="cri-o://6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75" gracePeriod=30 Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.339722 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5"] Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.339929 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" podUID="d7581c21-cae9-4c2b-8a46-3c08bcd54180" containerName="route-controller-manager" containerID="cri-o://e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497" gracePeriod=30 Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.629291 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.684029 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.732734 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-proxy-ca-bundles\") pod \"846186fd-c540-464d-b85a-a67f0c8870b6\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.732782 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-config\") pod \"846186fd-c540-464d-b85a-a67f0c8870b6\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.732803 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-client-ca\") pod \"846186fd-c540-464d-b85a-a67f0c8870b6\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.732877 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/846186fd-c540-464d-b85a-a67f0c8870b6-serving-cert\") pod \"846186fd-c540-464d-b85a-a67f0c8870b6\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.732910 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bqtk\" (UniqueName: \"kubernetes.io/projected/846186fd-c540-464d-b85a-a67f0c8870b6-kube-api-access-2bqtk\") pod \"846186fd-c540-464d-b85a-a67f0c8870b6\" (UID: \"846186fd-c540-464d-b85a-a67f0c8870b6\") " Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.734287 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "846186fd-c540-464d-b85a-a67f0c8870b6" (UID: "846186fd-c540-464d-b85a-a67f0c8870b6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.734345 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-config" (OuterVolumeSpecName: "config") pod "846186fd-c540-464d-b85a-a67f0c8870b6" (UID: "846186fd-c540-464d-b85a-a67f0c8870b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.734437 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-client-ca" (OuterVolumeSpecName: "client-ca") pod "846186fd-c540-464d-b85a-a67f0c8870b6" (UID: "846186fd-c540-464d-b85a-a67f0c8870b6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.739161 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/846186fd-c540-464d-b85a-a67f0c8870b6-kube-api-access-2bqtk" (OuterVolumeSpecName: "kube-api-access-2bqtk") pod "846186fd-c540-464d-b85a-a67f0c8870b6" (UID: "846186fd-c540-464d-b85a-a67f0c8870b6"). InnerVolumeSpecName "kube-api-access-2bqtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.739344 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846186fd-c540-464d-b85a-a67f0c8870b6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "846186fd-c540-464d-b85a-a67f0c8870b6" (UID: "846186fd-c540-464d-b85a-a67f0c8870b6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.812860 4969 generic.go:334] "Generic (PLEG): container finished" podID="d7581c21-cae9-4c2b-8a46-3c08bcd54180" containerID="e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497" exitCode=0 Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.812932 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" event={"ID":"d7581c21-cae9-4c2b-8a46-3c08bcd54180","Type":"ContainerDied","Data":"e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497"} Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.813034 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" event={"ID":"d7581c21-cae9-4c2b-8a46-3c08bcd54180","Type":"ContainerDied","Data":"1f4560b83a48b765e3585ff0eeb73b63063715c87d45f30a53142cad0917f303"} Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.812974 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.813106 4969 scope.go:117] "RemoveContainer" containerID="e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.814307 4969 generic.go:334] "Generic (PLEG): container finished" podID="846186fd-c540-464d-b85a-a67f0c8870b6" containerID="6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75" exitCode=0 Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.814344 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" event={"ID":"846186fd-c540-464d-b85a-a67f0c8870b6","Type":"ContainerDied","Data":"6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75"} Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.814371 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" event={"ID":"846186fd-c540-464d-b85a-a67f0c8870b6","Type":"ContainerDied","Data":"7d7c7e7c645a090f95d9cb1e9ccdd5c9dfc6d3582000563c1906ac2ef089f7e5"} Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.814487 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lrtsv" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.834163 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-client-ca\") pod \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.834262 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7581c21-cae9-4c2b-8a46-3c08bcd54180-serving-cert\") pod \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.834290 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jpw9\" (UniqueName: \"kubernetes.io/projected/d7581c21-cae9-4c2b-8a46-3c08bcd54180-kube-api-access-5jpw9\") pod \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.834447 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-config\") pod \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\" (UID: \"d7581c21-cae9-4c2b-8a46-3c08bcd54180\") " Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.834797 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/846186fd-c540-464d-b85a-a67f0c8870b6-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.834820 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bqtk\" (UniqueName: \"kubernetes.io/projected/846186fd-c540-464d-b85a-a67f0c8870b6-kube-api-access-2bqtk\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.834833 4969 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.834846 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.834856 4969 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/846186fd-c540-464d-b85a-a67f0c8870b6-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.835367 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-config" (OuterVolumeSpecName: "config") pod "d7581c21-cae9-4c2b-8a46-3c08bcd54180" (UID: "d7581c21-cae9-4c2b-8a46-3c08bcd54180"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.835402 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-client-ca" (OuterVolumeSpecName: "client-ca") pod "d7581c21-cae9-4c2b-8a46-3c08bcd54180" (UID: "d7581c21-cae9-4c2b-8a46-3c08bcd54180"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.837673 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7581c21-cae9-4c2b-8a46-3c08bcd54180-kube-api-access-5jpw9" (OuterVolumeSpecName: "kube-api-access-5jpw9") pod "d7581c21-cae9-4c2b-8a46-3c08bcd54180" (UID: "d7581c21-cae9-4c2b-8a46-3c08bcd54180"). InnerVolumeSpecName "kube-api-access-5jpw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.839181 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7581c21-cae9-4c2b-8a46-3c08bcd54180-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d7581c21-cae9-4c2b-8a46-3c08bcd54180" (UID: "d7581c21-cae9-4c2b-8a46-3c08bcd54180"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.839542 4969 scope.go:117] "RemoveContainer" containerID="e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497" Oct 04 08:28:42 crc kubenswrapper[4969]: E1004 08:28:42.839911 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497\": container with ID starting with e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497 not found: ID does not exist" containerID="e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.839957 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497"} err="failed to get container status \"e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497\": rpc error: code = NotFound desc = could not find container \"e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497\": container with ID starting with e4b8fc796db221fead59122b54ebfce9dd623d710f86a4f768179df04a8af497 not found: ID does not exist" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.839986 4969 scope.go:117] "RemoveContainer" containerID="6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.856745 4969 scope.go:117] "RemoveContainer" containerID="6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.857352 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lrtsv"] Oct 04 08:28:42 crc kubenswrapper[4969]: E1004 08:28:42.857451 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75\": container with ID starting with 6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75 not found: ID does not exist" containerID="6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.857618 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75"} err="failed to get container status \"6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75\": rpc error: code = NotFound desc = could not find container \"6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75\": container with ID starting with 6c303f6cb0fbac03147c17da6691d959daed70b01204fcf86a4a68da16de0f75 not found: ID does not exist" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.866401 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lrtsv"] Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.936125 4969 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.936173 4969 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7581c21-cae9-4c2b-8a46-3c08bcd54180-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.936195 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jpw9\" (UniqueName: \"kubernetes.io/projected/d7581c21-cae9-4c2b-8a46-3c08bcd54180-kube-api-access-5jpw9\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:42 crc kubenswrapper[4969]: I1004 08:28:42.936216 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7581c21-cae9-4c2b-8a46-3c08bcd54180-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:28:43 crc kubenswrapper[4969]: I1004 08:28:43.074316 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="846186fd-c540-464d-b85a-a67f0c8870b6" path="/var/lib/kubelet/pods/846186fd-c540-464d-b85a-a67f0c8870b6/volumes" Oct 04 08:28:43 crc kubenswrapper[4969]: I1004 08:28:43.139328 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5"] Oct 04 08:28:43 crc kubenswrapper[4969]: I1004 08:28:43.149676 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rshw5"] Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.141155 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6779b57b6-2qmgk"] Oct 04 08:28:44 crc kubenswrapper[4969]: E1004 08:28:44.141532 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7581c21-cae9-4c2b-8a46-3c08bcd54180" containerName="route-controller-manager" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.141553 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7581c21-cae9-4c2b-8a46-3c08bcd54180" containerName="route-controller-manager" Oct 04 08:28:44 crc kubenswrapper[4969]: E1004 08:28:44.141587 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="846186fd-c540-464d-b85a-a67f0c8870b6" containerName="controller-manager" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.141601 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="846186fd-c540-464d-b85a-a67f0c8870b6" containerName="controller-manager" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.141793 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7581c21-cae9-4c2b-8a46-3c08bcd54180" containerName="route-controller-manager" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.141828 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="846186fd-c540-464d-b85a-a67f0c8870b6" containerName="controller-manager" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.142447 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.145213 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.145694 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.146279 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.147278 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.147886 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.149593 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.155184 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.156985 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz"] Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.157811 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:44 crc kubenswrapper[4969]: W1004 08:28:44.160277 4969 reflector.go:561] object-"openshift-route-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 04 08:28:44 crc kubenswrapper[4969]: E1004 08:28:44.160595 4969 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 08:28:44 crc kubenswrapper[4969]: W1004 08:28:44.160281 4969 reflector.go:561] object-"openshift-route-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 04 08:28:44 crc kubenswrapper[4969]: E1004 08:28:44.160651 4969 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 08:28:44 crc kubenswrapper[4969]: W1004 08:28:44.162794 4969 reflector.go:561] object-"openshift-route-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 04 08:28:44 crc kubenswrapper[4969]: W1004 08:28:44.162901 4969 reflector.go:561] object-"openshift-route-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 04 08:28:44 crc kubenswrapper[4969]: E1004 08:28:44.162952 4969 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 08:28:44 crc kubenswrapper[4969]: W1004 08:28:44.162830 4969 reflector.go:561] object-"openshift-route-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 04 08:28:44 crc kubenswrapper[4969]: E1004 08:28:44.162987 4969 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 08:28:44 crc kubenswrapper[4969]: W1004 08:28:44.162838 4969 reflector.go:561] object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2": failed to list *v1.Secret: secrets "route-controller-manager-sa-dockercfg-h2zr2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.162860 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6779b57b6-2qmgk"] Oct 04 08:28:44 crc kubenswrapper[4969]: E1004 08:28:44.163011 4969 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-h2zr2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"route-controller-manager-sa-dockercfg-h2zr2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 08:28:44 crc kubenswrapper[4969]: E1004 08:28:44.163208 4969 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.172483 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz"] Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.252583 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzzl8\" (UniqueName: \"kubernetes.io/projected/adc06190-ad75-45ce-8cc3-28b267d76559-kube-api-access-pzzl8\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.252648 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d3d85bf-c86a-4b53-bdd6-736df1807422-client-ca\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.252673 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/adc06190-ad75-45ce-8cc3-28b267d76559-client-ca\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.252702 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc06190-ad75-45ce-8cc3-28b267d76559-config\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.252886 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d3d85bf-c86a-4b53-bdd6-736df1807422-serving-cert\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.252913 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5d3d85bf-c86a-4b53-bdd6-736df1807422-proxy-ca-bundles\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.252936 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsqlf\" (UniqueName: \"kubernetes.io/projected/5d3d85bf-c86a-4b53-bdd6-736df1807422-kube-api-access-vsqlf\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.252980 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adc06190-ad75-45ce-8cc3-28b267d76559-serving-cert\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.253003 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d3d85bf-c86a-4b53-bdd6-736df1807422-config\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.354268 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adc06190-ad75-45ce-8cc3-28b267d76559-serving-cert\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.354361 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d3d85bf-c86a-4b53-bdd6-736df1807422-config\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.354416 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzzl8\" (UniqueName: \"kubernetes.io/projected/adc06190-ad75-45ce-8cc3-28b267d76559-kube-api-access-pzzl8\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.354564 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d3d85bf-c86a-4b53-bdd6-736df1807422-client-ca\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.354602 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/adc06190-ad75-45ce-8cc3-28b267d76559-client-ca\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.354644 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc06190-ad75-45ce-8cc3-28b267d76559-config\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.354709 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d3d85bf-c86a-4b53-bdd6-736df1807422-serving-cert\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.354753 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5d3d85bf-c86a-4b53-bdd6-736df1807422-proxy-ca-bundles\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.354792 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsqlf\" (UniqueName: \"kubernetes.io/projected/5d3d85bf-c86a-4b53-bdd6-736df1807422-kube-api-access-vsqlf\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.355764 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d3d85bf-c86a-4b53-bdd6-736df1807422-config\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.356042 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d3d85bf-c86a-4b53-bdd6-736df1807422-client-ca\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.356318 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5d3d85bf-c86a-4b53-bdd6-736df1807422-proxy-ca-bundles\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.363910 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d3d85bf-c86a-4b53-bdd6-736df1807422-serving-cert\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.382299 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsqlf\" (UniqueName: \"kubernetes.io/projected/5d3d85bf-c86a-4b53-bdd6-736df1807422-kube-api-access-vsqlf\") pod \"controller-manager-6779b57b6-2qmgk\" (UID: \"5d3d85bf-c86a-4b53-bdd6-736df1807422\") " pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.459851 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.677455 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6779b57b6-2qmgk"] Oct 04 08:28:44 crc kubenswrapper[4969]: W1004 08:28:44.683352 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d3d85bf_c86a_4b53_bdd6_736df1807422.slice/crio-2606fd6a99576ced37d32f2fe9587bd7a8da48954a5e6d1aa795197e354642ea WatchSource:0}: Error finding container 2606fd6a99576ced37d32f2fe9587bd7a8da48954a5e6d1aa795197e354642ea: Status 404 returned error can't find the container with id 2606fd6a99576ced37d32f2fe9587bd7a8da48954a5e6d1aa795197e354642ea Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.838399 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" event={"ID":"5d3d85bf-c86a-4b53-bdd6-736df1807422","Type":"ContainerStarted","Data":"88c4646880f611f319d98f9c98973fd979219f44e98fbcd34fb5fde8f66e59cb"} Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.838500 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" event={"ID":"5d3d85bf-c86a-4b53-bdd6-736df1807422","Type":"ContainerStarted","Data":"2606fd6a99576ced37d32f2fe9587bd7a8da48954a5e6d1aa795197e354642ea"} Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.838814 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.842973 4969 patch_prober.go:28] interesting pod/controller-manager-6779b57b6-2qmgk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" start-of-body= Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.843059 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" podUID="5d3d85bf-c86a-4b53-bdd6-736df1807422" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" Oct 04 08:28:44 crc kubenswrapper[4969]: I1004 08:28:44.874123 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" podStartSLOduration=0.874096023 podStartE2EDuration="874.096023ms" podCreationTimestamp="2025-10-04 08:28:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:28:44.865352515 +0000 UTC m=+752.619621359" watchObservedRunningTime="2025-10-04 08:28:44.874096023 +0000 UTC m=+752.628364887" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.057108 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.059763 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.063932 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7581c21-cae9-4c2b-8a46-3c08bcd54180" path="/var/lib/kubelet/pods/d7581c21-cae9-4c2b-8a46-3c08bcd54180/volumes" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.065876 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/adc06190-ad75-45ce-8cc3-28b267d76559-client-ca\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.095495 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.107013 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc06190-ad75-45ce-8cc3-28b267d76559-config\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.315158 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.331985 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzzl8\" (UniqueName: \"kubernetes.io/projected/adc06190-ad75-45ce-8cc3-28b267d76559-kube-api-access-pzzl8\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:45 crc kubenswrapper[4969]: E1004 08:28:45.355123 4969 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 08:28:45 crc kubenswrapper[4969]: E1004 08:28:45.355211 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/adc06190-ad75-45ce-8cc3-28b267d76559-serving-cert podName:adc06190-ad75-45ce-8cc3-28b267d76559 nodeName:}" failed. No retries permitted until 2025-10-04 08:28:45.855191659 +0000 UTC m=+753.609460473 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/adc06190-ad75-45ce-8cc3-28b267d76559-serving-cert") pod "route-controller-manager-595bdb4cf8-5shkz" (UID: "adc06190-ad75-45ce-8cc3-28b267d76559") : failed to sync secret cache: timed out waiting for the condition Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.367497 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.715339 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.851172 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6779b57b6-2qmgk" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.876028 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adc06190-ad75-45ce-8cc3-28b267d76559-serving-cert\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.881310 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adc06190-ad75-45ce-8cc3-28b267d76559-serving-cert\") pod \"route-controller-manager-595bdb4cf8-5shkz\" (UID: \"adc06190-ad75-45ce-8cc3-28b267d76559\") " pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:45 crc kubenswrapper[4969]: I1004 08:28:45.971453 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.413208 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz"] Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.806031 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh"] Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.807405 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.810670 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-t29j9" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.830723 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh"] Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.853004 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-5mjb6"] Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.857375 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.861062 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd"] Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.861768 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" event={"ID":"adc06190-ad75-45ce-8cc3-28b267d76559","Type":"ContainerStarted","Data":"05e37494fccf1c58668343fe640df8b013c50f1cc47cbd9a808133821535add7"} Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.861796 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" event={"ID":"adc06190-ad75-45ce-8cc3-28b267d76559","Type":"ContainerStarted","Data":"c8956ff0b2e56523cf8b40f369a5713adb8f7d38912778534e50e920d3e03216"} Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.861861 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.872814 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.878388 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd"] Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.888800 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99nkg\" (UniqueName: \"kubernetes.io/projected/af2281a5-7ed5-4d7b-b1ab-f97948b7fd59-kube-api-access-99nkg\") pod \"nmstate-metrics-fdff9cb8d-t54hh\" (UID: \"af2281a5-7ed5-4d7b-b1ab-f97948b7fd59\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.926410 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" podStartSLOduration=2.926394263 podStartE2EDuration="2.926394263s" podCreationTimestamp="2025-10-04 08:28:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:28:46.923231585 +0000 UTC m=+754.677500399" watchObservedRunningTime="2025-10-04 08:28:46.926394263 +0000 UTC m=+754.680663077" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.954749 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l"] Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.955555 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.962069 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-cj7dj" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.963840 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.969167 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.978195 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l"] Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.990050 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/dcb7ec91-3c05-474d-aead-c3014ca7e646-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-rsbrd\" (UID: \"dcb7ec91-3c05-474d-aead-c3014ca7e646\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.990095 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l752j\" (UniqueName: \"kubernetes.io/projected/dcb7ec91-3c05-474d-aead-c3014ca7e646-kube-api-access-l752j\") pod \"nmstate-webhook-6cdbc54649-rsbrd\" (UID: \"dcb7ec91-3c05-474d-aead-c3014ca7e646\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.990120 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-ovs-socket\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.990183 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-dbus-socket\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.990209 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cj6k\" (UniqueName: \"kubernetes.io/projected/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-kube-api-access-7cj6k\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.990250 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99nkg\" (UniqueName: \"kubernetes.io/projected/af2281a5-7ed5-4d7b-b1ab-f97948b7fd59-kube-api-access-99nkg\") pod \"nmstate-metrics-fdff9cb8d-t54hh\" (UID: \"af2281a5-7ed5-4d7b-b1ab-f97948b7fd59\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh" Oct 04 08:28:46 crc kubenswrapper[4969]: I1004 08:28:46.990281 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-nmstate-lock\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.009433 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99nkg\" (UniqueName: \"kubernetes.io/projected/af2281a5-7ed5-4d7b-b1ab-f97948b7fd59-kube-api-access-99nkg\") pod \"nmstate-metrics-fdff9cb8d-t54hh\" (UID: \"af2281a5-7ed5-4d7b-b1ab-f97948b7fd59\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.091249 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/dcb7ec91-3c05-474d-aead-c3014ca7e646-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-rsbrd\" (UID: \"dcb7ec91-3c05-474d-aead-c3014ca7e646\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.091443 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l752j\" (UniqueName: \"kubernetes.io/projected/dcb7ec91-3c05-474d-aead-c3014ca7e646-kube-api-access-l752j\") pod \"nmstate-webhook-6cdbc54649-rsbrd\" (UID: \"dcb7ec91-3c05-474d-aead-c3014ca7e646\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.091558 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-ovs-socket\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.091654 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-dbus-socket\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.091691 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-ovs-socket\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.091782 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-mnl4l\" (UID: \"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.091870 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cj6k\" (UniqueName: \"kubernetes.io/projected/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-kube-api-access-7cj6k\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.091951 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-mnl4l\" (UID: \"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.092028 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-nmstate-lock\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.091985 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-dbus-socket\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.092149 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld9l4\" (UniqueName: \"kubernetes.io/projected/7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43-kube-api-access-ld9l4\") pod \"nmstate-console-plugin-6b874cbd85-mnl4l\" (UID: \"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.092214 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-nmstate-lock\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.095266 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/dcb7ec91-3c05-474d-aead-c3014ca7e646-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-rsbrd\" (UID: \"dcb7ec91-3c05-474d-aead-c3014ca7e646\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.111262 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l752j\" (UniqueName: \"kubernetes.io/projected/dcb7ec91-3c05-474d-aead-c3014ca7e646-kube-api-access-l752j\") pod \"nmstate-webhook-6cdbc54649-rsbrd\" (UID: \"dcb7ec91-3c05-474d-aead-c3014ca7e646\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.114994 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cj6k\" (UniqueName: \"kubernetes.io/projected/eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9-kube-api-access-7cj6k\") pod \"nmstate-handler-5mjb6\" (UID: \"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9\") " pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.155326 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6bdc9cf6df-v8hnw"] Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.156000 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.158205 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.188951 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.193263 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld9l4\" (UniqueName: \"kubernetes.io/projected/7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43-kube-api-access-ld9l4\") pod \"nmstate-console-plugin-6b874cbd85-mnl4l\" (UID: \"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.193352 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-mnl4l\" (UID: \"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.193386 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-mnl4l\" (UID: \"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.194462 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-mnl4l\" (UID: \"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.198928 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-mnl4l\" (UID: \"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.201837 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.233365 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld9l4\" (UniqueName: \"kubernetes.io/projected/7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43-kube-api-access-ld9l4\") pod \"nmstate-console-plugin-6b874cbd85-mnl4l\" (UID: \"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.241889 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6bdc9cf6df-v8hnw"] Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.269898 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.297002 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1852544c-0a51-455c-bc2d-631e4eb0f748-console-oauth-config\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.297063 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-trusted-ca-bundle\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.297092 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-oauth-serving-cert\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.297113 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1852544c-0a51-455c-bc2d-631e4eb0f748-console-serving-cert\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.297138 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grv8q\" (UniqueName: \"kubernetes.io/projected/1852544c-0a51-455c-bc2d-631e4eb0f748-kube-api-access-grv8q\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.297157 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-console-config\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.297179 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-service-ca\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.398085 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grv8q\" (UniqueName: \"kubernetes.io/projected/1852544c-0a51-455c-bc2d-631e4eb0f748-kube-api-access-grv8q\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.398130 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-console-config\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.398160 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-service-ca\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.398193 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1852544c-0a51-455c-bc2d-631e4eb0f748-console-oauth-config\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.398228 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-trusted-ca-bundle\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.398259 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-oauth-serving-cert\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.398284 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1852544c-0a51-455c-bc2d-631e4eb0f748-console-serving-cert\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.399560 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-trusted-ca-bundle\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.399581 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-console-config\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.399736 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-oauth-serving-cert\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.400072 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1852544c-0a51-455c-bc2d-631e4eb0f748-service-ca\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.403440 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1852544c-0a51-455c-bc2d-631e4eb0f748-console-oauth-config\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.403952 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1852544c-0a51-455c-bc2d-631e4eb0f748-console-serving-cert\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.449454 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grv8q\" (UniqueName: \"kubernetes.io/projected/1852544c-0a51-455c-bc2d-631e4eb0f748-kube-api-access-grv8q\") pod \"console-6bdc9cf6df-v8hnw\" (UID: \"1852544c-0a51-455c-bc2d-631e4eb0f748\") " pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.469171 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.606382 4969 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.725612 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh"] Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.851773 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l"] Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.857308 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd"] Oct 04 08:28:47 crc kubenswrapper[4969]: W1004 08:28:47.860323 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f4625c5_29dd_4cf4_8bb6_9bc88fe16a43.slice/crio-b998d5af228d93f6826ea4eaf90096e7e4e834d4ef7592ff05316d77a991dcf1 WatchSource:0}: Error finding container b998d5af228d93f6826ea4eaf90096e7e4e834d4ef7592ff05316d77a991dcf1: Status 404 returned error can't find the container with id b998d5af228d93f6826ea4eaf90096e7e4e834d4ef7592ff05316d77a991dcf1 Oct 04 08:28:47 crc kubenswrapper[4969]: W1004 08:28:47.861042 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcb7ec91_3c05_474d_aead_c3014ca7e646.slice/crio-b9859d14b6de4325dc2511cfeb69d4205ec89e44b5cc090bb88fb7a88e9b038c WatchSource:0}: Error finding container b9859d14b6de4325dc2511cfeb69d4205ec89e44b5cc090bb88fb7a88e9b038c: Status 404 returned error can't find the container with id b9859d14b6de4325dc2511cfeb69d4205ec89e44b5cc090bb88fb7a88e9b038c Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.875941 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5mjb6" event={"ID":"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9","Type":"ContainerStarted","Data":"ee94f8ca545c9f82aa325b5a52ab6ee6e94f25c68ca5b39f4072e055249afadb"} Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.877262 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh" event={"ID":"af2281a5-7ed5-4d7b-b1ab-f97948b7fd59","Type":"ContainerStarted","Data":"a8673f5114c031ae90c750eff3bd0da0b191b63c21723118a179a7b3aedb5197"} Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.878775 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" event={"ID":"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43","Type":"ContainerStarted","Data":"b998d5af228d93f6826ea4eaf90096e7e4e834d4ef7592ff05316d77a991dcf1"} Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.879159 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:47 crc kubenswrapper[4969]: I1004 08:28:47.885162 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-595bdb4cf8-5shkz" Oct 04 08:28:48 crc kubenswrapper[4969]: I1004 08:28:48.055260 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6bdc9cf6df-v8hnw"] Oct 04 08:28:48 crc kubenswrapper[4969]: I1004 08:28:48.887922 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" event={"ID":"dcb7ec91-3c05-474d-aead-c3014ca7e646","Type":"ContainerStarted","Data":"b9859d14b6de4325dc2511cfeb69d4205ec89e44b5cc090bb88fb7a88e9b038c"} Oct 04 08:28:48 crc kubenswrapper[4969]: I1004 08:28:48.889581 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6bdc9cf6df-v8hnw" event={"ID":"1852544c-0a51-455c-bc2d-631e4eb0f748","Type":"ContainerStarted","Data":"b5cc6a81f76a7c4db9b5d87dd38daeaa84512b871085288225650a930b4bfc7a"} Oct 04 08:28:48 crc kubenswrapper[4969]: I1004 08:28:48.889632 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6bdc9cf6df-v8hnw" event={"ID":"1852544c-0a51-455c-bc2d-631e4eb0f748","Type":"ContainerStarted","Data":"63edd4359c17d2540c988f0721f4fe985f0cc43cca90dffb99576ffbc62d9303"} Oct 04 08:28:48 crc kubenswrapper[4969]: I1004 08:28:48.915197 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6bdc9cf6df-v8hnw" podStartSLOduration=1.915178922 podStartE2EDuration="1.915178922s" podCreationTimestamp="2025-10-04 08:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:28:48.911586783 +0000 UTC m=+756.665855647" watchObservedRunningTime="2025-10-04 08:28:48.915178922 +0000 UTC m=+756.669447756" Oct 04 08:28:49 crc kubenswrapper[4969]: I1004 08:28:49.666854 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:28:49 crc kubenswrapper[4969]: I1004 08:28:49.667148 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:28:51 crc kubenswrapper[4969]: I1004 08:28:51.916969 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh" event={"ID":"af2281a5-7ed5-4d7b-b1ab-f97948b7fd59","Type":"ContainerStarted","Data":"e171bc2c270f1b16b1d523302d5dcf23bf9bde10c03116c6fc5eae5afc485eb7"} Oct 04 08:28:51 crc kubenswrapper[4969]: I1004 08:28:51.918442 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" event={"ID":"7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43","Type":"ContainerStarted","Data":"0657a156193ecdaee3e67fd44f9f9ed87cca70e5049f885ea41af4cd33a31e7d"} Oct 04 08:28:51 crc kubenswrapper[4969]: I1004 08:28:51.920659 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5mjb6" event={"ID":"eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9","Type":"ContainerStarted","Data":"a81a6e17c62033a0dff20e45cc853e5d0e919cd055bb4112c1e17abc48bebec8"} Oct 04 08:28:51 crc kubenswrapper[4969]: I1004 08:28:51.920740 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:51 crc kubenswrapper[4969]: I1004 08:28:51.922295 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" event={"ID":"dcb7ec91-3c05-474d-aead-c3014ca7e646","Type":"ContainerStarted","Data":"a7e5732314da9738c1847e4bcf72c3b58cf2e29649dfc54fbd36eb7a8e410832"} Oct 04 08:28:51 crc kubenswrapper[4969]: I1004 08:28:51.922411 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:28:51 crc kubenswrapper[4969]: I1004 08:28:51.938637 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mnl4l" podStartSLOduration=2.721708325 podStartE2EDuration="5.938619478s" podCreationTimestamp="2025-10-04 08:28:46 +0000 UTC" firstStartedPulling="2025-10-04 08:28:47.863949372 +0000 UTC m=+755.618218206" lastFinishedPulling="2025-10-04 08:28:51.080860545 +0000 UTC m=+758.835129359" observedRunningTime="2025-10-04 08:28:51.935985063 +0000 UTC m=+759.690253887" watchObservedRunningTime="2025-10-04 08:28:51.938619478 +0000 UTC m=+759.692888292" Oct 04 08:28:51 crc kubenswrapper[4969]: I1004 08:28:51.957890 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-5mjb6" podStartSLOduration=2.092669126 podStartE2EDuration="5.957867357s" podCreationTimestamp="2025-10-04 08:28:46 +0000 UTC" firstStartedPulling="2025-10-04 08:28:47.227721134 +0000 UTC m=+754.981989948" lastFinishedPulling="2025-10-04 08:28:51.092919365 +0000 UTC m=+758.847188179" observedRunningTime="2025-10-04 08:28:51.951405737 +0000 UTC m=+759.705674591" watchObservedRunningTime="2025-10-04 08:28:51.957867357 +0000 UTC m=+759.712136191" Oct 04 08:28:52 crc kubenswrapper[4969]: I1004 08:28:52.019052 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" podStartSLOduration=2.779636339 podStartE2EDuration="6.01902379s" podCreationTimestamp="2025-10-04 08:28:46 +0000 UTC" firstStartedPulling="2025-10-04 08:28:47.869661635 +0000 UTC m=+755.623930459" lastFinishedPulling="2025-10-04 08:28:51.109049096 +0000 UTC m=+758.863317910" observedRunningTime="2025-10-04 08:28:52.015396349 +0000 UTC m=+759.769665163" watchObservedRunningTime="2025-10-04 08:28:52.01902379 +0000 UTC m=+759.773292624" Oct 04 08:28:53 crc kubenswrapper[4969]: I1004 08:28:53.939001 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh" event={"ID":"af2281a5-7ed5-4d7b-b1ab-f97948b7fd59","Type":"ContainerStarted","Data":"37ba49e8f1b18d8fd383a0e93369ce87ef7a8907da2f335b98cd879aa470eeaa"} Oct 04 08:28:53 crc kubenswrapper[4969]: I1004 08:28:53.966796 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-t54hh" podStartSLOduration=2.028400686 podStartE2EDuration="7.966777087s" podCreationTimestamp="2025-10-04 08:28:46 +0000 UTC" firstStartedPulling="2025-10-04 08:28:47.742479479 +0000 UTC m=+755.496748293" lastFinishedPulling="2025-10-04 08:28:53.68085588 +0000 UTC m=+761.435124694" observedRunningTime="2025-10-04 08:28:53.963229219 +0000 UTC m=+761.717498083" watchObservedRunningTime="2025-10-04 08:28:53.966777087 +0000 UTC m=+761.721045911" Oct 04 08:28:57 crc kubenswrapper[4969]: I1004 08:28:57.217692 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-5mjb6" Oct 04 08:28:57 crc kubenswrapper[4969]: I1004 08:28:57.469975 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:57 crc kubenswrapper[4969]: I1004 08:28:57.470034 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:57 crc kubenswrapper[4969]: I1004 08:28:57.476835 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:57 crc kubenswrapper[4969]: I1004 08:28:57.974972 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6bdc9cf6df-v8hnw" Oct 04 08:28:58 crc kubenswrapper[4969]: I1004 08:28:58.044121 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m67tq"] Oct 04 08:29:07 crc kubenswrapper[4969]: I1004 08:29:07.211944 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rsbrd" Oct 04 08:29:19 crc kubenswrapper[4969]: I1004 08:29:19.666575 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:29:19 crc kubenswrapper[4969]: I1004 08:29:19.667570 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:29:19 crc kubenswrapper[4969]: I1004 08:29:19.667637 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:29:19 crc kubenswrapper[4969]: I1004 08:29:19.668534 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"47fb93c2519799da0fd36fa2c011f7f4604debf093aa2814af4cd13eb886ba6f"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:29:19 crc kubenswrapper[4969]: I1004 08:29:19.668630 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://47fb93c2519799da0fd36fa2c011f7f4604debf093aa2814af4cd13eb886ba6f" gracePeriod=600 Oct 04 08:29:20 crc kubenswrapper[4969]: I1004 08:29:20.149505 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="47fb93c2519799da0fd36fa2c011f7f4604debf093aa2814af4cd13eb886ba6f" exitCode=0 Oct 04 08:29:20 crc kubenswrapper[4969]: I1004 08:29:20.149563 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"47fb93c2519799da0fd36fa2c011f7f4604debf093aa2814af4cd13eb886ba6f"} Oct 04 08:29:20 crc kubenswrapper[4969]: I1004 08:29:20.150016 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"a0e7ef25f0eb92534c8c59058c5d643ea03dee7ad9ad7534679a9b23c47e914c"} Oct 04 08:29:20 crc kubenswrapper[4969]: I1004 08:29:20.150036 4969 scope.go:117] "RemoveContainer" containerID="89de65150d90f75568f74a82023f8dca9943d3891a3087d9e41cf1dc1b47d259" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.106320 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-m67tq" podUID="1a654328-5c34-4927-949f-0b5818da9b49" containerName="console" containerID="cri-o://90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39" gracePeriod=15 Oct 04 08:29:23 crc kubenswrapper[4969]: E1004 08:29:23.214293 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a654328_5c34_4927_949f_0b5818da9b49.slice/crio-90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a654328_5c34_4927_949f_0b5818da9b49.slice/crio-conmon-90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39.scope\": RecentStats: unable to find data in memory cache]" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.623602 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m67tq_1a654328-5c34-4927-949f-0b5818da9b49/console/0.log" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.623945 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.751830 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-serving-cert\") pod \"1a654328-5c34-4927-949f-0b5818da9b49\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.752247 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-console-config\") pod \"1a654328-5c34-4927-949f-0b5818da9b49\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.752325 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9ffd\" (UniqueName: \"kubernetes.io/projected/1a654328-5c34-4927-949f-0b5818da9b49-kube-api-access-g9ffd\") pod \"1a654328-5c34-4927-949f-0b5818da9b49\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.752454 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-oauth-serving-cert\") pod \"1a654328-5c34-4927-949f-0b5818da9b49\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.752540 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-trusted-ca-bundle\") pod \"1a654328-5c34-4927-949f-0b5818da9b49\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.752574 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-oauth-config\") pod \"1a654328-5c34-4927-949f-0b5818da9b49\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.752626 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-service-ca\") pod \"1a654328-5c34-4927-949f-0b5818da9b49\" (UID: \"1a654328-5c34-4927-949f-0b5818da9b49\") " Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.753247 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-console-config" (OuterVolumeSpecName: "console-config") pod "1a654328-5c34-4927-949f-0b5818da9b49" (UID: "1a654328-5c34-4927-949f-0b5818da9b49"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.753784 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-service-ca" (OuterVolumeSpecName: "service-ca") pod "1a654328-5c34-4927-949f-0b5818da9b49" (UID: "1a654328-5c34-4927-949f-0b5818da9b49"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.753778 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "1a654328-5c34-4927-949f-0b5818da9b49" (UID: "1a654328-5c34-4927-949f-0b5818da9b49"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.753854 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1a654328-5c34-4927-949f-0b5818da9b49" (UID: "1a654328-5c34-4927-949f-0b5818da9b49"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.758843 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "1a654328-5c34-4927-949f-0b5818da9b49" (UID: "1a654328-5c34-4927-949f-0b5818da9b49"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.759486 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "1a654328-5c34-4927-949f-0b5818da9b49" (UID: "1a654328-5c34-4927-949f-0b5818da9b49"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.759788 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a654328-5c34-4927-949f-0b5818da9b49-kube-api-access-g9ffd" (OuterVolumeSpecName: "kube-api-access-g9ffd") pod "1a654328-5c34-4927-949f-0b5818da9b49" (UID: "1a654328-5c34-4927-949f-0b5818da9b49"). InnerVolumeSpecName "kube-api-access-g9ffd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.854842 4969 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.854887 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9ffd\" (UniqueName: \"kubernetes.io/projected/1a654328-5c34-4927-949f-0b5818da9b49-kube-api-access-g9ffd\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.854903 4969 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.854914 4969 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.854925 4969 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.854937 4969 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1a654328-5c34-4927-949f-0b5818da9b49-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:23 crc kubenswrapper[4969]: I1004 08:29:23.854948 4969 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a654328-5c34-4927-949f-0b5818da9b49-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.185749 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m67tq_1a654328-5c34-4927-949f-0b5818da9b49/console/0.log" Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.185817 4969 generic.go:334] "Generic (PLEG): container finished" podID="1a654328-5c34-4927-949f-0b5818da9b49" containerID="90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39" exitCode=2 Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.185868 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m67tq" event={"ID":"1a654328-5c34-4927-949f-0b5818da9b49","Type":"ContainerDied","Data":"90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39"} Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.185911 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m67tq" event={"ID":"1a654328-5c34-4927-949f-0b5818da9b49","Type":"ContainerDied","Data":"b36c16b280146cdb6ba3655fb90d48769c9017b4363badd79ceae11424cd653b"} Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.185907 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m67tq" Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.185932 4969 scope.go:117] "RemoveContainer" containerID="90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39" Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.224082 4969 scope.go:117] "RemoveContainer" containerID="90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39" Oct 04 08:29:24 crc kubenswrapper[4969]: E1004 08:29:24.225153 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39\": container with ID starting with 90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39 not found: ID does not exist" containerID="90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39" Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.225194 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39"} err="failed to get container status \"90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39\": rpc error: code = NotFound desc = could not find container \"90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39\": container with ID starting with 90de51a9fa78b00c49dc47a2824c9e3f769329575efde1f588c6287d0d8a0f39 not found: ID does not exist" Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.231109 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m67tq"] Oct 04 08:29:24 crc kubenswrapper[4969]: I1004 08:29:24.231158 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-m67tq"] Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.067661 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a654328-5c34-4927-949f-0b5818da9b49" path="/var/lib/kubelet/pods/1a654328-5c34-4927-949f-0b5818da9b49/volumes" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.182375 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5"] Oct 04 08:29:25 crc kubenswrapper[4969]: E1004 08:29:25.182669 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a654328-5c34-4927-949f-0b5818da9b49" containerName="console" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.182691 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a654328-5c34-4927-949f-0b5818da9b49" containerName="console" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.182829 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a654328-5c34-4927-949f-0b5818da9b49" containerName="console" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.185490 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.187838 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.204151 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5"] Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.270738 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5gwm\" (UniqueName: \"kubernetes.io/projected/71587e21-d80e-404c-8ccf-aff572933817-kube-api-access-c5gwm\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.270786 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.271126 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.372125 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.372197 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.372234 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5gwm\" (UniqueName: \"kubernetes.io/projected/71587e21-d80e-404c-8ccf-aff572933817-kube-api-access-c5gwm\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.373198 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.373581 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.398934 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5gwm\" (UniqueName: \"kubernetes.io/projected/71587e21-d80e-404c-8ccf-aff572933817-kube-api-access-c5gwm\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.505006 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:25 crc kubenswrapper[4969]: I1004 08:29:25.770016 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5"] Oct 04 08:29:26 crc kubenswrapper[4969]: I1004 08:29:26.203125 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" event={"ID":"71587e21-d80e-404c-8ccf-aff572933817","Type":"ContainerStarted","Data":"e7deecbcbf0d0ff6f96fceb4b94a8e8261d2caef4001f2ec1807ac62064815b1"} Oct 04 08:29:26 crc kubenswrapper[4969]: I1004 08:29:26.204312 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" event={"ID":"71587e21-d80e-404c-8ccf-aff572933817","Type":"ContainerStarted","Data":"42c1fece59632c81db8282e0247f67a61fbe56ecfce7e05b2102d699657208e8"} Oct 04 08:29:27 crc kubenswrapper[4969]: I1004 08:29:27.217902 4969 generic.go:334] "Generic (PLEG): container finished" podID="71587e21-d80e-404c-8ccf-aff572933817" containerID="e7deecbcbf0d0ff6f96fceb4b94a8e8261d2caef4001f2ec1807ac62064815b1" exitCode=0 Oct 04 08:29:27 crc kubenswrapper[4969]: I1004 08:29:27.218346 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" event={"ID":"71587e21-d80e-404c-8ccf-aff572933817","Type":"ContainerDied","Data":"e7deecbcbf0d0ff6f96fceb4b94a8e8261d2caef4001f2ec1807ac62064815b1"} Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.697297 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ppghl"] Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.698950 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.724971 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ppghl"] Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.821521 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-catalog-content\") pod \"redhat-operators-ppghl\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.821591 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-utilities\") pod \"redhat-operators-ppghl\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.821632 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f84kp\" (UniqueName: \"kubernetes.io/projected/674e9e49-5ac8-482e-b370-e72593958989-kube-api-access-f84kp\") pod \"redhat-operators-ppghl\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.923476 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-catalog-content\") pod \"redhat-operators-ppghl\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.923557 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-utilities\") pod \"redhat-operators-ppghl\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.923598 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f84kp\" (UniqueName: \"kubernetes.io/projected/674e9e49-5ac8-482e-b370-e72593958989-kube-api-access-f84kp\") pod \"redhat-operators-ppghl\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.924071 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-catalog-content\") pod \"redhat-operators-ppghl\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.924068 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-utilities\") pod \"redhat-operators-ppghl\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:28 crc kubenswrapper[4969]: I1004 08:29:28.943782 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f84kp\" (UniqueName: \"kubernetes.io/projected/674e9e49-5ac8-482e-b370-e72593958989-kube-api-access-f84kp\") pod \"redhat-operators-ppghl\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:29 crc kubenswrapper[4969]: I1004 08:29:29.017837 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:29 crc kubenswrapper[4969]: I1004 08:29:29.232969 4969 generic.go:334] "Generic (PLEG): container finished" podID="71587e21-d80e-404c-8ccf-aff572933817" containerID="22308ce4b0e722d9fd68203886db957b0880234f7c7eb6c6f37db0bb540a640f" exitCode=0 Oct 04 08:29:29 crc kubenswrapper[4969]: I1004 08:29:29.233040 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" event={"ID":"71587e21-d80e-404c-8ccf-aff572933817","Type":"ContainerDied","Data":"22308ce4b0e722d9fd68203886db957b0880234f7c7eb6c6f37db0bb540a640f"} Oct 04 08:29:29 crc kubenswrapper[4969]: I1004 08:29:29.490619 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ppghl"] Oct 04 08:29:30 crc kubenswrapper[4969]: I1004 08:29:30.243654 4969 generic.go:334] "Generic (PLEG): container finished" podID="71587e21-d80e-404c-8ccf-aff572933817" containerID="49e7bb9063405a4596e5c474a90c738d9d7a92e2e04bef46630c10653f88897e" exitCode=0 Oct 04 08:29:30 crc kubenswrapper[4969]: I1004 08:29:30.243767 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" event={"ID":"71587e21-d80e-404c-8ccf-aff572933817","Type":"ContainerDied","Data":"49e7bb9063405a4596e5c474a90c738d9d7a92e2e04bef46630c10653f88897e"} Oct 04 08:29:30 crc kubenswrapper[4969]: I1004 08:29:30.246019 4969 generic.go:334] "Generic (PLEG): container finished" podID="674e9e49-5ac8-482e-b370-e72593958989" containerID="7a5b8cf10dd559976dab343e0c4ae03a7284cbd6ebe40aa4bfea2ec00b742f0e" exitCode=0 Oct 04 08:29:30 crc kubenswrapper[4969]: I1004 08:29:30.246087 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppghl" event={"ID":"674e9e49-5ac8-482e-b370-e72593958989","Type":"ContainerDied","Data":"7a5b8cf10dd559976dab343e0c4ae03a7284cbd6ebe40aa4bfea2ec00b742f0e"} Oct 04 08:29:30 crc kubenswrapper[4969]: I1004 08:29:30.246127 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppghl" event={"ID":"674e9e49-5ac8-482e-b370-e72593958989","Type":"ContainerStarted","Data":"248a1df5ffa5730b5ef9a8a8cc52c4c1afadda9eeeecb983359d51e407f0e0e8"} Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.257035 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppghl" event={"ID":"674e9e49-5ac8-482e-b370-e72593958989","Type":"ContainerStarted","Data":"e4535fcf3d624fe09e6e713c0af4145f16363b67a088728f11cf501e85364344"} Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.582158 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.673978 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5gwm\" (UniqueName: \"kubernetes.io/projected/71587e21-d80e-404c-8ccf-aff572933817-kube-api-access-c5gwm\") pod \"71587e21-d80e-404c-8ccf-aff572933817\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.674088 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-bundle\") pod \"71587e21-d80e-404c-8ccf-aff572933817\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.674109 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-util\") pod \"71587e21-d80e-404c-8ccf-aff572933817\" (UID: \"71587e21-d80e-404c-8ccf-aff572933817\") " Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.675378 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-bundle" (OuterVolumeSpecName: "bundle") pod "71587e21-d80e-404c-8ccf-aff572933817" (UID: "71587e21-d80e-404c-8ccf-aff572933817"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.688686 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71587e21-d80e-404c-8ccf-aff572933817-kube-api-access-c5gwm" (OuterVolumeSpecName: "kube-api-access-c5gwm") pod "71587e21-d80e-404c-8ccf-aff572933817" (UID: "71587e21-d80e-404c-8ccf-aff572933817"). InnerVolumeSpecName "kube-api-access-c5gwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.775368 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5gwm\" (UniqueName: \"kubernetes.io/projected/71587e21-d80e-404c-8ccf-aff572933817-kube-api-access-c5gwm\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.775397 4969 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.936822 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-util" (OuterVolumeSpecName: "util") pod "71587e21-d80e-404c-8ccf-aff572933817" (UID: "71587e21-d80e-404c-8ccf-aff572933817"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:29:31 crc kubenswrapper[4969]: I1004 08:29:31.984398 4969 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71587e21-d80e-404c-8ccf-aff572933817-util\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:32 crc kubenswrapper[4969]: I1004 08:29:32.268818 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" event={"ID":"71587e21-d80e-404c-8ccf-aff572933817","Type":"ContainerDied","Data":"42c1fece59632c81db8282e0247f67a61fbe56ecfce7e05b2102d699657208e8"} Oct 04 08:29:32 crc kubenswrapper[4969]: I1004 08:29:32.268891 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42c1fece59632c81db8282e0247f67a61fbe56ecfce7e05b2102d699657208e8" Oct 04 08:29:32 crc kubenswrapper[4969]: I1004 08:29:32.268840 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5" Oct 04 08:29:32 crc kubenswrapper[4969]: I1004 08:29:32.276827 4969 generic.go:334] "Generic (PLEG): container finished" podID="674e9e49-5ac8-482e-b370-e72593958989" containerID="e4535fcf3d624fe09e6e713c0af4145f16363b67a088728f11cf501e85364344" exitCode=0 Oct 04 08:29:32 crc kubenswrapper[4969]: I1004 08:29:32.276881 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppghl" event={"ID":"674e9e49-5ac8-482e-b370-e72593958989","Type":"ContainerDied","Data":"e4535fcf3d624fe09e6e713c0af4145f16363b67a088728f11cf501e85364344"} Oct 04 08:29:33 crc kubenswrapper[4969]: I1004 08:29:33.286560 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppghl" event={"ID":"674e9e49-5ac8-482e-b370-e72593958989","Type":"ContainerStarted","Data":"f7110b9be954d061b372cc369ea067745842e30ebb2e03329fe98b60a1cd22bf"} Oct 04 08:29:39 crc kubenswrapper[4969]: I1004 08:29:39.018031 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:39 crc kubenswrapper[4969]: I1004 08:29:39.018562 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:39 crc kubenswrapper[4969]: I1004 08:29:39.068017 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:39 crc kubenswrapper[4969]: I1004 08:29:39.091640 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ppghl" podStartSLOduration=8.600195361 podStartE2EDuration="11.091620352s" podCreationTimestamp="2025-10-04 08:29:28 +0000 UTC" firstStartedPulling="2025-10-04 08:29:30.2492572 +0000 UTC m=+798.003526044" lastFinishedPulling="2025-10-04 08:29:32.740682211 +0000 UTC m=+800.494951035" observedRunningTime="2025-10-04 08:29:33.314103556 +0000 UTC m=+801.068372400" watchObservedRunningTime="2025-10-04 08:29:39.091620352 +0000 UTC m=+806.845889176" Oct 04 08:29:39 crc kubenswrapper[4969]: I1004 08:29:39.364631 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.458171 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n"] Oct 04 08:29:42 crc kubenswrapper[4969]: E1004 08:29:42.458378 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71587e21-d80e-404c-8ccf-aff572933817" containerName="extract" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.458389 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="71587e21-d80e-404c-8ccf-aff572933817" containerName="extract" Oct 04 08:29:42 crc kubenswrapper[4969]: E1004 08:29:42.458399 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71587e21-d80e-404c-8ccf-aff572933817" containerName="pull" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.458406 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="71587e21-d80e-404c-8ccf-aff572933817" containerName="pull" Oct 04 08:29:42 crc kubenswrapper[4969]: E1004 08:29:42.458428 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71587e21-d80e-404c-8ccf-aff572933817" containerName="util" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.458434 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="71587e21-d80e-404c-8ccf-aff572933817" containerName="util" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.472606 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="71587e21-d80e-404c-8ccf-aff572933817" containerName="extract" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.473369 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.475867 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.476375 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-4zgwj" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.480801 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.481986 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.488469 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n"] Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.492348 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.632401 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/67b0a26f-7d5e-4988-83ba-8073480c2e2c-webhook-cert\") pod \"metallb-operator-controller-manager-74c8fbc96b-hr67n\" (UID: \"67b0a26f-7d5e-4988-83ba-8073480c2e2c\") " pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.632571 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz2mk\" (UniqueName: \"kubernetes.io/projected/67b0a26f-7d5e-4988-83ba-8073480c2e2c-kube-api-access-qz2mk\") pod \"metallb-operator-controller-manager-74c8fbc96b-hr67n\" (UID: \"67b0a26f-7d5e-4988-83ba-8073480c2e2c\") " pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.632765 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/67b0a26f-7d5e-4988-83ba-8073480c2e2c-apiservice-cert\") pod \"metallb-operator-controller-manager-74c8fbc96b-hr67n\" (UID: \"67b0a26f-7d5e-4988-83ba-8073480c2e2c\") " pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.686851 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ppghl"] Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.687068 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ppghl" podUID="674e9e49-5ac8-482e-b370-e72593958989" containerName="registry-server" containerID="cri-o://f7110b9be954d061b372cc369ea067745842e30ebb2e03329fe98b60a1cd22bf" gracePeriod=2 Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.733500 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/67b0a26f-7d5e-4988-83ba-8073480c2e2c-apiservice-cert\") pod \"metallb-operator-controller-manager-74c8fbc96b-hr67n\" (UID: \"67b0a26f-7d5e-4988-83ba-8073480c2e2c\") " pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.733610 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/67b0a26f-7d5e-4988-83ba-8073480c2e2c-webhook-cert\") pod \"metallb-operator-controller-manager-74c8fbc96b-hr67n\" (UID: \"67b0a26f-7d5e-4988-83ba-8073480c2e2c\") " pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.733639 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz2mk\" (UniqueName: \"kubernetes.io/projected/67b0a26f-7d5e-4988-83ba-8073480c2e2c-kube-api-access-qz2mk\") pod \"metallb-operator-controller-manager-74c8fbc96b-hr67n\" (UID: \"67b0a26f-7d5e-4988-83ba-8073480c2e2c\") " pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.741656 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/67b0a26f-7d5e-4988-83ba-8073480c2e2c-apiservice-cert\") pod \"metallb-operator-controller-manager-74c8fbc96b-hr67n\" (UID: \"67b0a26f-7d5e-4988-83ba-8073480c2e2c\") " pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.745793 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/67b0a26f-7d5e-4988-83ba-8073480c2e2c-webhook-cert\") pod \"metallb-operator-controller-manager-74c8fbc96b-hr67n\" (UID: \"67b0a26f-7d5e-4988-83ba-8073480c2e2c\") " pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.750855 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz2mk\" (UniqueName: \"kubernetes.io/projected/67b0a26f-7d5e-4988-83ba-8073480c2e2c-kube-api-access-qz2mk\") pod \"metallb-operator-controller-manager-74c8fbc96b-hr67n\" (UID: \"67b0a26f-7d5e-4988-83ba-8073480c2e2c\") " pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.781460 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk"] Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.782303 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.784623 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.784670 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.784623 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-854lw" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.791764 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.805507 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk"] Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.938410 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d0240b5e-b8bb-4690-a95a-f291b834c0d0-webhook-cert\") pod \"metallb-operator-webhook-server-6bd7b55ffd-6gshk\" (UID: \"d0240b5e-b8bb-4690-a95a-f291b834c0d0\") " pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.938991 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz6lx\" (UniqueName: \"kubernetes.io/projected/d0240b5e-b8bb-4690-a95a-f291b834c0d0-kube-api-access-zz6lx\") pod \"metallb-operator-webhook-server-6bd7b55ffd-6gshk\" (UID: \"d0240b5e-b8bb-4690-a95a-f291b834c0d0\") " pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:42 crc kubenswrapper[4969]: I1004 08:29:42.939065 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d0240b5e-b8bb-4690-a95a-f291b834c0d0-apiservice-cert\") pod \"metallb-operator-webhook-server-6bd7b55ffd-6gshk\" (UID: \"d0240b5e-b8bb-4690-a95a-f291b834c0d0\") " pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.046082 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d0240b5e-b8bb-4690-a95a-f291b834c0d0-webhook-cert\") pod \"metallb-operator-webhook-server-6bd7b55ffd-6gshk\" (UID: \"d0240b5e-b8bb-4690-a95a-f291b834c0d0\") " pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.046137 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz6lx\" (UniqueName: \"kubernetes.io/projected/d0240b5e-b8bb-4690-a95a-f291b834c0d0-kube-api-access-zz6lx\") pod \"metallb-operator-webhook-server-6bd7b55ffd-6gshk\" (UID: \"d0240b5e-b8bb-4690-a95a-f291b834c0d0\") " pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.046201 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d0240b5e-b8bb-4690-a95a-f291b834c0d0-apiservice-cert\") pod \"metallb-operator-webhook-server-6bd7b55ffd-6gshk\" (UID: \"d0240b5e-b8bb-4690-a95a-f291b834c0d0\") " pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.060261 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d0240b5e-b8bb-4690-a95a-f291b834c0d0-webhook-cert\") pod \"metallb-operator-webhook-server-6bd7b55ffd-6gshk\" (UID: \"d0240b5e-b8bb-4690-a95a-f291b834c0d0\") " pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.063166 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d0240b5e-b8bb-4690-a95a-f291b834c0d0-apiservice-cert\") pod \"metallb-operator-webhook-server-6bd7b55ffd-6gshk\" (UID: \"d0240b5e-b8bb-4690-a95a-f291b834c0d0\") " pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.073075 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz6lx\" (UniqueName: \"kubernetes.io/projected/d0240b5e-b8bb-4690-a95a-f291b834c0d0-kube-api-access-zz6lx\") pod \"metallb-operator-webhook-server-6bd7b55ffd-6gshk\" (UID: \"d0240b5e-b8bb-4690-a95a-f291b834c0d0\") " pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.116781 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n"] Oct 04 08:29:43 crc kubenswrapper[4969]: W1004 08:29:43.128610 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67b0a26f_7d5e_4988_83ba_8073480c2e2c.slice/crio-e198729b1bbd81afce3712015a1a9e9d7d93b7e635c5968f0afc61d2367ed856 WatchSource:0}: Error finding container e198729b1bbd81afce3712015a1a9e9d7d93b7e635c5968f0afc61d2367ed856: Status 404 returned error can't find the container with id e198729b1bbd81afce3712015a1a9e9d7d93b7e635c5968f0afc61d2367ed856 Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.204734 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.357930 4969 generic.go:334] "Generic (PLEG): container finished" podID="674e9e49-5ac8-482e-b370-e72593958989" containerID="f7110b9be954d061b372cc369ea067745842e30ebb2e03329fe98b60a1cd22bf" exitCode=0 Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.357980 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppghl" event={"ID":"674e9e49-5ac8-482e-b370-e72593958989","Type":"ContainerDied","Data":"f7110b9be954d061b372cc369ea067745842e30ebb2e03329fe98b60a1cd22bf"} Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.358762 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" event={"ID":"67b0a26f-7d5e-4988-83ba-8073480c2e2c","Type":"ContainerStarted","Data":"e198729b1bbd81afce3712015a1a9e9d7d93b7e635c5968f0afc61d2367ed856"} Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.479577 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk"] Oct 04 08:29:43 crc kubenswrapper[4969]: W1004 08:29:43.504300 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0240b5e_b8bb_4690_a95a_f291b834c0d0.slice/crio-5a3c5dd2414b732f2058c993d340e32db92685c780e5e163f2ad2bb1445ac0f4 WatchSource:0}: Error finding container 5a3c5dd2414b732f2058c993d340e32db92685c780e5e163f2ad2bb1445ac0f4: Status 404 returned error can't find the container with id 5a3c5dd2414b732f2058c993d340e32db92685c780e5e163f2ad2bb1445ac0f4 Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.609130 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.754820 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f84kp\" (UniqueName: \"kubernetes.io/projected/674e9e49-5ac8-482e-b370-e72593958989-kube-api-access-f84kp\") pod \"674e9e49-5ac8-482e-b370-e72593958989\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.755161 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-utilities\") pod \"674e9e49-5ac8-482e-b370-e72593958989\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.755193 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-catalog-content\") pod \"674e9e49-5ac8-482e-b370-e72593958989\" (UID: \"674e9e49-5ac8-482e-b370-e72593958989\") " Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.755973 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-utilities" (OuterVolumeSpecName: "utilities") pod "674e9e49-5ac8-482e-b370-e72593958989" (UID: "674e9e49-5ac8-482e-b370-e72593958989"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.762807 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/674e9e49-5ac8-482e-b370-e72593958989-kube-api-access-f84kp" (OuterVolumeSpecName: "kube-api-access-f84kp") pod "674e9e49-5ac8-482e-b370-e72593958989" (UID: "674e9e49-5ac8-482e-b370-e72593958989"). InnerVolumeSpecName "kube-api-access-f84kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.838688 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "674e9e49-5ac8-482e-b370-e72593958989" (UID: "674e9e49-5ac8-482e-b370-e72593958989"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.856850 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.856894 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f84kp\" (UniqueName: \"kubernetes.io/projected/674e9e49-5ac8-482e-b370-e72593958989-kube-api-access-f84kp\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:43 crc kubenswrapper[4969]: I1004 08:29:43.856907 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674e9e49-5ac8-482e-b370-e72593958989-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:29:44 crc kubenswrapper[4969]: I1004 08:29:44.370001 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" event={"ID":"d0240b5e-b8bb-4690-a95a-f291b834c0d0","Type":"ContainerStarted","Data":"5a3c5dd2414b732f2058c993d340e32db92685c780e5e163f2ad2bb1445ac0f4"} Oct 04 08:29:44 crc kubenswrapper[4969]: I1004 08:29:44.373292 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppghl" event={"ID":"674e9e49-5ac8-482e-b370-e72593958989","Type":"ContainerDied","Data":"248a1df5ffa5730b5ef9a8a8cc52c4c1afadda9eeeecb983359d51e407f0e0e8"} Oct 04 08:29:44 crc kubenswrapper[4969]: I1004 08:29:44.373326 4969 scope.go:117] "RemoveContainer" containerID="f7110b9be954d061b372cc369ea067745842e30ebb2e03329fe98b60a1cd22bf" Oct 04 08:29:44 crc kubenswrapper[4969]: I1004 08:29:44.373486 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppghl" Oct 04 08:29:44 crc kubenswrapper[4969]: I1004 08:29:44.404279 4969 scope.go:117] "RemoveContainer" containerID="e4535fcf3d624fe09e6e713c0af4145f16363b67a088728f11cf501e85364344" Oct 04 08:29:44 crc kubenswrapper[4969]: I1004 08:29:44.406841 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ppghl"] Oct 04 08:29:44 crc kubenswrapper[4969]: I1004 08:29:44.412261 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ppghl"] Oct 04 08:29:44 crc kubenswrapper[4969]: I1004 08:29:44.417853 4969 scope.go:117] "RemoveContainer" containerID="7a5b8cf10dd559976dab343e0c4ae03a7284cbd6ebe40aa4bfea2ec00b742f0e" Oct 04 08:29:45 crc kubenswrapper[4969]: I1004 08:29:45.073635 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="674e9e49-5ac8-482e-b370-e72593958989" path="/var/lib/kubelet/pods/674e9e49-5ac8-482e-b370-e72593958989/volumes" Oct 04 08:29:46 crc kubenswrapper[4969]: I1004 08:29:46.404050 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" event={"ID":"67b0a26f-7d5e-4988-83ba-8073480c2e2c","Type":"ContainerStarted","Data":"82ce61ff37a75b88fba81e15412f9cd20c5e83b5bab2e07efce1f298b8b505b8"} Oct 04 08:29:46 crc kubenswrapper[4969]: I1004 08:29:46.404214 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:29:46 crc kubenswrapper[4969]: I1004 08:29:46.425677 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" podStartSLOduration=1.59218009 podStartE2EDuration="4.425661387s" podCreationTimestamp="2025-10-04 08:29:42 +0000 UTC" firstStartedPulling="2025-10-04 08:29:43.133324557 +0000 UTC m=+810.887593371" lastFinishedPulling="2025-10-04 08:29:45.966805854 +0000 UTC m=+813.721074668" observedRunningTime="2025-10-04 08:29:46.418402946 +0000 UTC m=+814.172671750" watchObservedRunningTime="2025-10-04 08:29:46.425661387 +0000 UTC m=+814.179930201" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.416719 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" event={"ID":"d0240b5e-b8bb-4690-a95a-f291b834c0d0","Type":"ContainerStarted","Data":"4618318a8380287459f707db090029dbd5b528fd0a92c4c9f0c95b9f906032c1"} Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.417036 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.441389 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" podStartSLOduration=2.004046822 podStartE2EDuration="6.441371356s" podCreationTimestamp="2025-10-04 08:29:42 +0000 UTC" firstStartedPulling="2025-10-04 08:29:43.506951708 +0000 UTC m=+811.261220522" lastFinishedPulling="2025-10-04 08:29:47.944276242 +0000 UTC m=+815.698545056" observedRunningTime="2025-10-04 08:29:48.441070348 +0000 UTC m=+816.195339162" watchObservedRunningTime="2025-10-04 08:29:48.441371356 +0000 UTC m=+816.195640170" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.890916 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-clf9n"] Oct 04 08:29:48 crc kubenswrapper[4969]: E1004 08:29:48.891186 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="674e9e49-5ac8-482e-b370-e72593958989" containerName="extract-content" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.891200 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="674e9e49-5ac8-482e-b370-e72593958989" containerName="extract-content" Oct 04 08:29:48 crc kubenswrapper[4969]: E1004 08:29:48.891213 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="674e9e49-5ac8-482e-b370-e72593958989" containerName="extract-utilities" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.891221 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="674e9e49-5ac8-482e-b370-e72593958989" containerName="extract-utilities" Oct 04 08:29:48 crc kubenswrapper[4969]: E1004 08:29:48.891236 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="674e9e49-5ac8-482e-b370-e72593958989" containerName="registry-server" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.891244 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="674e9e49-5ac8-482e-b370-e72593958989" containerName="registry-server" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.891382 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="674e9e49-5ac8-482e-b370-e72593958989" containerName="registry-server" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.892383 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.915394 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-clf9n"] Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.930982 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tk2q\" (UniqueName: \"kubernetes.io/projected/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-kube-api-access-2tk2q\") pod \"certified-operators-clf9n\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.931091 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-catalog-content\") pod \"certified-operators-clf9n\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:48 crc kubenswrapper[4969]: I1004 08:29:48.931147 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-utilities\") pod \"certified-operators-clf9n\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:49 crc kubenswrapper[4969]: I1004 08:29:49.032398 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tk2q\" (UniqueName: \"kubernetes.io/projected/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-kube-api-access-2tk2q\") pod \"certified-operators-clf9n\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:49 crc kubenswrapper[4969]: I1004 08:29:49.032805 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-catalog-content\") pod \"certified-operators-clf9n\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:49 crc kubenswrapper[4969]: I1004 08:29:49.032844 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-utilities\") pod \"certified-operators-clf9n\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:49 crc kubenswrapper[4969]: I1004 08:29:49.033253 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-catalog-content\") pod \"certified-operators-clf9n\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:49 crc kubenswrapper[4969]: I1004 08:29:49.033401 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-utilities\") pod \"certified-operators-clf9n\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:49 crc kubenswrapper[4969]: I1004 08:29:49.062802 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tk2q\" (UniqueName: \"kubernetes.io/projected/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-kube-api-access-2tk2q\") pod \"certified-operators-clf9n\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:49 crc kubenswrapper[4969]: I1004 08:29:49.207769 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:49 crc kubenswrapper[4969]: I1004 08:29:49.638857 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-clf9n"] Oct 04 08:29:50 crc kubenswrapper[4969]: I1004 08:29:50.443495 4969 generic.go:334] "Generic (PLEG): container finished" podID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerID="1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e" exitCode=0 Oct 04 08:29:50 crc kubenswrapper[4969]: I1004 08:29:50.443557 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clf9n" event={"ID":"23b0e0cd-ffee-4f95-818a-089c6ffd48ad","Type":"ContainerDied","Data":"1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e"} Oct 04 08:29:50 crc kubenswrapper[4969]: I1004 08:29:50.443588 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clf9n" event={"ID":"23b0e0cd-ffee-4f95-818a-089c6ffd48ad","Type":"ContainerStarted","Data":"d95fbd21fc9ed6fcf7b4af6544c309bc57a6ca331b1efcbfe1a533038b4752d8"} Oct 04 08:29:52 crc kubenswrapper[4969]: I1004 08:29:52.460123 4969 generic.go:334] "Generic (PLEG): container finished" podID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerID="474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45" exitCode=0 Oct 04 08:29:52 crc kubenswrapper[4969]: I1004 08:29:52.460184 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clf9n" event={"ID":"23b0e0cd-ffee-4f95-818a-089c6ffd48ad","Type":"ContainerDied","Data":"474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45"} Oct 04 08:29:53 crc kubenswrapper[4969]: I1004 08:29:53.477583 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clf9n" event={"ID":"23b0e0cd-ffee-4f95-818a-089c6ffd48ad","Type":"ContainerStarted","Data":"4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d"} Oct 04 08:29:53 crc kubenswrapper[4969]: I1004 08:29:53.500746 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-clf9n" podStartSLOduration=3.038850389 podStartE2EDuration="5.500733134s" podCreationTimestamp="2025-10-04 08:29:48 +0000 UTC" firstStartedPulling="2025-10-04 08:29:50.445234621 +0000 UTC m=+818.199503425" lastFinishedPulling="2025-10-04 08:29:52.907117336 +0000 UTC m=+820.661386170" observedRunningTime="2025-10-04 08:29:53.49978061 +0000 UTC m=+821.254049424" watchObservedRunningTime="2025-10-04 08:29:53.500733134 +0000 UTC m=+821.255001948" Oct 04 08:29:59 crc kubenswrapper[4969]: I1004 08:29:59.208655 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:59 crc kubenswrapper[4969]: I1004 08:29:59.209149 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:59 crc kubenswrapper[4969]: I1004 08:29:59.259633 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:29:59 crc kubenswrapper[4969]: I1004 08:29:59.566749 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.142520 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs"] Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.143695 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.146828 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.147084 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.166487 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs"] Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.196635 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c57b46a1-a938-4dc6-99c0-23c044e307da-secret-volume\") pod \"collect-profiles-29326110-txvxs\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.196694 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-224th\" (UniqueName: \"kubernetes.io/projected/c57b46a1-a938-4dc6-99c0-23c044e307da-kube-api-access-224th\") pod \"collect-profiles-29326110-txvxs\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.196804 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c57b46a1-a938-4dc6-99c0-23c044e307da-config-volume\") pod \"collect-profiles-29326110-txvxs\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.298190 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c57b46a1-a938-4dc6-99c0-23c044e307da-secret-volume\") pod \"collect-profiles-29326110-txvxs\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.298234 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-224th\" (UniqueName: \"kubernetes.io/projected/c57b46a1-a938-4dc6-99c0-23c044e307da-kube-api-access-224th\") pod \"collect-profiles-29326110-txvxs\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.298279 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c57b46a1-a938-4dc6-99c0-23c044e307da-config-volume\") pod \"collect-profiles-29326110-txvxs\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.299076 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c57b46a1-a938-4dc6-99c0-23c044e307da-config-volume\") pod \"collect-profiles-29326110-txvxs\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.304086 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c57b46a1-a938-4dc6-99c0-23c044e307da-secret-volume\") pod \"collect-profiles-29326110-txvxs\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.319219 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-224th\" (UniqueName: \"kubernetes.io/projected/c57b46a1-a938-4dc6-99c0-23c044e307da-kube-api-access-224th\") pod \"collect-profiles-29326110-txvxs\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.468217 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:00 crc kubenswrapper[4969]: I1004 08:30:00.926373 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs"] Oct 04 08:30:01 crc kubenswrapper[4969]: I1004 08:30:01.532373 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" event={"ID":"c57b46a1-a938-4dc6-99c0-23c044e307da","Type":"ContainerDied","Data":"237c72f9d595bc3676d27b336cb77671d0283fe37474435a4c51d760fcf105c4"} Oct 04 08:30:01 crc kubenswrapper[4969]: I1004 08:30:01.532408 4969 generic.go:334] "Generic (PLEG): container finished" podID="c57b46a1-a938-4dc6-99c0-23c044e307da" containerID="237c72f9d595bc3676d27b336cb77671d0283fe37474435a4c51d760fcf105c4" exitCode=0 Oct 04 08:30:01 crc kubenswrapper[4969]: I1004 08:30:01.532485 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" event={"ID":"c57b46a1-a938-4dc6-99c0-23c044e307da","Type":"ContainerStarted","Data":"c3e3b8e600177341b82d119fcf6dff20bdc46601f326fc8c291c042b63a1ba92"} Oct 04 08:30:01 crc kubenswrapper[4969]: I1004 08:30:01.686665 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-clf9n"] Oct 04 08:30:02 crc kubenswrapper[4969]: I1004 08:30:02.540944 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-clf9n" podUID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerName="registry-server" containerID="cri-o://4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d" gracePeriod=2 Oct 04 08:30:02 crc kubenswrapper[4969]: I1004 08:30:02.934889 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:02 crc kubenswrapper[4969]: I1004 08:30:02.937643 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c57b46a1-a938-4dc6-99c0-23c044e307da-secret-volume\") pod \"c57b46a1-a938-4dc6-99c0-23c044e307da\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " Oct 04 08:30:02 crc kubenswrapper[4969]: I1004 08:30:02.937694 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-224th\" (UniqueName: \"kubernetes.io/projected/c57b46a1-a938-4dc6-99c0-23c044e307da-kube-api-access-224th\") pod \"c57b46a1-a938-4dc6-99c0-23c044e307da\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " Oct 04 08:30:02 crc kubenswrapper[4969]: I1004 08:30:02.937737 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c57b46a1-a938-4dc6-99c0-23c044e307da-config-volume\") pod \"c57b46a1-a938-4dc6-99c0-23c044e307da\" (UID: \"c57b46a1-a938-4dc6-99c0-23c044e307da\") " Oct 04 08:30:02 crc kubenswrapper[4969]: I1004 08:30:02.938465 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c57b46a1-a938-4dc6-99c0-23c044e307da-config-volume" (OuterVolumeSpecName: "config-volume") pod "c57b46a1-a938-4dc6-99c0-23c044e307da" (UID: "c57b46a1-a938-4dc6-99c0-23c044e307da"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:30:02 crc kubenswrapper[4969]: I1004 08:30:02.943267 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c57b46a1-a938-4dc6-99c0-23c044e307da-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c57b46a1-a938-4dc6-99c0-23c044e307da" (UID: "c57b46a1-a938-4dc6-99c0-23c044e307da"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:30:02 crc kubenswrapper[4969]: I1004 08:30:02.975665 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c57b46a1-a938-4dc6-99c0-23c044e307da-kube-api-access-224th" (OuterVolumeSpecName: "kube-api-access-224th") pod "c57b46a1-a938-4dc6-99c0-23c044e307da" (UID: "c57b46a1-a938-4dc6-99c0-23c044e307da"). InnerVolumeSpecName "kube-api-access-224th". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.016136 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.039070 4969 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c57b46a1-a938-4dc6-99c0-23c044e307da-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.039106 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-224th\" (UniqueName: \"kubernetes.io/projected/c57b46a1-a938-4dc6-99c0-23c044e307da-kube-api-access-224th\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.039118 4969 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c57b46a1-a938-4dc6-99c0-23c044e307da-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.140873 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-catalog-content\") pod \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.141080 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tk2q\" (UniqueName: \"kubernetes.io/projected/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-kube-api-access-2tk2q\") pod \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.141205 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-utilities\") pod \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\" (UID: \"23b0e0cd-ffee-4f95-818a-089c6ffd48ad\") " Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.141925 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-utilities" (OuterVolumeSpecName: "utilities") pod "23b0e0cd-ffee-4f95-818a-089c6ffd48ad" (UID: "23b0e0cd-ffee-4f95-818a-089c6ffd48ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.142036 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.144026 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-kube-api-access-2tk2q" (OuterVolumeSpecName: "kube-api-access-2tk2q") pod "23b0e0cd-ffee-4f95-818a-089c6ffd48ad" (UID: "23b0e0cd-ffee-4f95-818a-089c6ffd48ad"). InnerVolumeSpecName "kube-api-access-2tk2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.184946 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23b0e0cd-ffee-4f95-818a-089c6ffd48ad" (UID: "23b0e0cd-ffee-4f95-818a-089c6ffd48ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.211752 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6bd7b55ffd-6gshk" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.242826 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.242863 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tk2q\" (UniqueName: \"kubernetes.io/projected/23b0e0cd-ffee-4f95-818a-089c6ffd48ad-kube-api-access-2tk2q\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.549407 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.549403 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs" event={"ID":"c57b46a1-a938-4dc6-99c0-23c044e307da","Type":"ContainerDied","Data":"c3e3b8e600177341b82d119fcf6dff20bdc46601f326fc8c291c042b63a1ba92"} Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.549630 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3e3b8e600177341b82d119fcf6dff20bdc46601f326fc8c291c042b63a1ba92" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.552977 4969 generic.go:334] "Generic (PLEG): container finished" podID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerID="4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d" exitCode=0 Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.553681 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clf9n" event={"ID":"23b0e0cd-ffee-4f95-818a-089c6ffd48ad","Type":"ContainerDied","Data":"4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d"} Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.553818 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clf9n" event={"ID":"23b0e0cd-ffee-4f95-818a-089c6ffd48ad","Type":"ContainerDied","Data":"d95fbd21fc9ed6fcf7b4af6544c309bc57a6ca331b1efcbfe1a533038b4752d8"} Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.553887 4969 scope.go:117] "RemoveContainer" containerID="4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.553714 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clf9n" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.577696 4969 scope.go:117] "RemoveContainer" containerID="474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.583368 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-clf9n"] Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.587700 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-clf9n"] Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.609834 4969 scope.go:117] "RemoveContainer" containerID="1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.626544 4969 scope.go:117] "RemoveContainer" containerID="4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d" Oct 04 08:30:03 crc kubenswrapper[4969]: E1004 08:30:03.627068 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d\": container with ID starting with 4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d not found: ID does not exist" containerID="4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.627127 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d"} err="failed to get container status \"4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d\": rpc error: code = NotFound desc = could not find container \"4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d\": container with ID starting with 4c4cefdd8783258d3abaefa724844dbf2047c99b49cca4bed396175e789e814d not found: ID does not exist" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.627195 4969 scope.go:117] "RemoveContainer" containerID="474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45" Oct 04 08:30:03 crc kubenswrapper[4969]: E1004 08:30:03.627584 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45\": container with ID starting with 474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45 not found: ID does not exist" containerID="474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.627611 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45"} err="failed to get container status \"474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45\": rpc error: code = NotFound desc = could not find container \"474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45\": container with ID starting with 474fcbd76b6f2d8f2ed7829b31d245b2ce0d899ed37b87954b8861f5e695aa45 not found: ID does not exist" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.627629 4969 scope.go:117] "RemoveContainer" containerID="1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e" Oct 04 08:30:03 crc kubenswrapper[4969]: E1004 08:30:03.627823 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e\": container with ID starting with 1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e not found: ID does not exist" containerID="1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e" Oct 04 08:30:03 crc kubenswrapper[4969]: I1004 08:30:03.627862 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e"} err="failed to get container status \"1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e\": rpc error: code = NotFound desc = could not find container \"1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e\": container with ID starting with 1ba9146525c28af908011b9892276f075b20d8d6b0335d6840bdd06da474435e not found: ID does not exist" Oct 04 08:30:05 crc kubenswrapper[4969]: I1004 08:30:05.063664 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" path="/var/lib/kubelet/pods/23b0e0cd-ffee-4f95-818a-089c6ffd48ad/volumes" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.303958 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfn5"] Oct 04 08:30:07 crc kubenswrapper[4969]: E1004 08:30:07.304317 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c57b46a1-a938-4dc6-99c0-23c044e307da" containerName="collect-profiles" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.304339 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c57b46a1-a938-4dc6-99c0-23c044e307da" containerName="collect-profiles" Oct 04 08:30:07 crc kubenswrapper[4969]: E1004 08:30:07.304360 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerName="extract-utilities" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.304374 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerName="extract-utilities" Oct 04 08:30:07 crc kubenswrapper[4969]: E1004 08:30:07.304403 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerName="extract-content" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.304449 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerName="extract-content" Oct 04 08:30:07 crc kubenswrapper[4969]: E1004 08:30:07.304477 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerName="registry-server" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.304490 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerName="registry-server" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.304778 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="23b0e0cd-ffee-4f95-818a-089c6ffd48ad" containerName="registry-server" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.304804 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c57b46a1-a938-4dc6-99c0-23c044e307da" containerName="collect-profiles" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.306266 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.321205 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfn5"] Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.393011 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-utilities\") pod \"redhat-marketplace-mlfn5\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.393058 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-catalog-content\") pod \"redhat-marketplace-mlfn5\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.393087 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96m55\" (UniqueName: \"kubernetes.io/projected/1aaa07d5-45f2-4074-af25-15ae361ff89e-kube-api-access-96m55\") pod \"redhat-marketplace-mlfn5\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.494251 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-catalog-content\") pod \"redhat-marketplace-mlfn5\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.494655 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96m55\" (UniqueName: \"kubernetes.io/projected/1aaa07d5-45f2-4074-af25-15ae361ff89e-kube-api-access-96m55\") pod \"redhat-marketplace-mlfn5\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.494793 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-utilities\") pod \"redhat-marketplace-mlfn5\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.495152 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-catalog-content\") pod \"redhat-marketplace-mlfn5\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.495329 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-utilities\") pod \"redhat-marketplace-mlfn5\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.519582 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96m55\" (UniqueName: \"kubernetes.io/projected/1aaa07d5-45f2-4074-af25-15ae361ff89e-kube-api-access-96m55\") pod \"redhat-marketplace-mlfn5\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:07 crc kubenswrapper[4969]: I1004 08:30:07.641661 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:08 crc kubenswrapper[4969]: I1004 08:30:08.086213 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfn5"] Oct 04 08:30:08 crc kubenswrapper[4969]: I1004 08:30:08.588510 4969 generic.go:334] "Generic (PLEG): container finished" podID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerID="3f2a0c9d900fa83797a047787121e6773a927ec043610b8d8e9c2594cf67d237" exitCode=0 Oct 04 08:30:08 crc kubenswrapper[4969]: I1004 08:30:08.588547 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfn5" event={"ID":"1aaa07d5-45f2-4074-af25-15ae361ff89e","Type":"ContainerDied","Data":"3f2a0c9d900fa83797a047787121e6773a927ec043610b8d8e9c2594cf67d237"} Oct 04 08:30:08 crc kubenswrapper[4969]: I1004 08:30:08.588569 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfn5" event={"ID":"1aaa07d5-45f2-4074-af25-15ae361ff89e","Type":"ContainerStarted","Data":"65749c01f5e044b49e4f6693e144173f822b46e1ed9670fbf14f845db87474c5"} Oct 04 08:30:09 crc kubenswrapper[4969]: I1004 08:30:09.596991 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfn5" event={"ID":"1aaa07d5-45f2-4074-af25-15ae361ff89e","Type":"ContainerStarted","Data":"b0cddc45a729fbee29630d542e4f295b87b2ff70a3f2bb09831b96d871c3d669"} Oct 04 08:30:10 crc kubenswrapper[4969]: I1004 08:30:10.604846 4969 generic.go:334] "Generic (PLEG): container finished" podID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerID="b0cddc45a729fbee29630d542e4f295b87b2ff70a3f2bb09831b96d871c3d669" exitCode=0 Oct 04 08:30:10 crc kubenswrapper[4969]: I1004 08:30:10.604905 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfn5" event={"ID":"1aaa07d5-45f2-4074-af25-15ae361ff89e","Type":"ContainerDied","Data":"b0cddc45a729fbee29630d542e4f295b87b2ff70a3f2bb09831b96d871c3d669"} Oct 04 08:30:11 crc kubenswrapper[4969]: I1004 08:30:11.614557 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfn5" event={"ID":"1aaa07d5-45f2-4074-af25-15ae361ff89e","Type":"ContainerStarted","Data":"f8c9bda401d4ff35c2e1d5075f74099768d5bcec416d35389e01c82115be5e09"} Oct 04 08:30:11 crc kubenswrapper[4969]: I1004 08:30:11.633034 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mlfn5" podStartSLOduration=2.215915331 podStartE2EDuration="4.633015992s" podCreationTimestamp="2025-10-04 08:30:07 +0000 UTC" firstStartedPulling="2025-10-04 08:30:08.589777204 +0000 UTC m=+836.344046028" lastFinishedPulling="2025-10-04 08:30:11.006877855 +0000 UTC m=+838.761146689" observedRunningTime="2025-10-04 08:30:11.629493855 +0000 UTC m=+839.383762679" watchObservedRunningTime="2025-10-04 08:30:11.633015992 +0000 UTC m=+839.387284826" Oct 04 08:30:17 crc kubenswrapper[4969]: I1004 08:30:17.643123 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:17 crc kubenswrapper[4969]: I1004 08:30:17.643974 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:17 crc kubenswrapper[4969]: I1004 08:30:17.702252 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:17 crc kubenswrapper[4969]: I1004 08:30:17.760942 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.107021 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tnvjn"] Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.109820 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.125716 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnvjn"] Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.262541 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-utilities\") pod \"community-operators-tnvjn\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.262665 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvlb7\" (UniqueName: \"kubernetes.io/projected/2333a480-04ca-4f93-937f-c91bf30323c2-kube-api-access-qvlb7\") pod \"community-operators-tnvjn\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.262869 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-catalog-content\") pod \"community-operators-tnvjn\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.364260 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-utilities\") pod \"community-operators-tnvjn\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.364334 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvlb7\" (UniqueName: \"kubernetes.io/projected/2333a480-04ca-4f93-937f-c91bf30323c2-kube-api-access-qvlb7\") pod \"community-operators-tnvjn\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.364379 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-catalog-content\") pod \"community-operators-tnvjn\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.364893 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-utilities\") pod \"community-operators-tnvjn\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.364921 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-catalog-content\") pod \"community-operators-tnvjn\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.383207 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvlb7\" (UniqueName: \"kubernetes.io/projected/2333a480-04ca-4f93-937f-c91bf30323c2-kube-api-access-qvlb7\") pod \"community-operators-tnvjn\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.439050 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:19 crc kubenswrapper[4969]: I1004 08:30:19.902903 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnvjn"] Oct 04 08:30:20 crc kubenswrapper[4969]: I1004 08:30:20.682517 4969 generic.go:334] "Generic (PLEG): container finished" podID="2333a480-04ca-4f93-937f-c91bf30323c2" containerID="3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f" exitCode=0 Oct 04 08:30:20 crc kubenswrapper[4969]: I1004 08:30:20.682589 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvjn" event={"ID":"2333a480-04ca-4f93-937f-c91bf30323c2","Type":"ContainerDied","Data":"3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f"} Oct 04 08:30:20 crc kubenswrapper[4969]: I1004 08:30:20.682631 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvjn" event={"ID":"2333a480-04ca-4f93-937f-c91bf30323c2","Type":"ContainerStarted","Data":"c3f8a50ebbcf8240f86b3d87b1a1403c4d42761f76026466b36c9244aaf63da3"} Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.289955 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfn5"] Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.290535 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mlfn5" podUID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerName="registry-server" containerID="cri-o://f8c9bda401d4ff35c2e1d5075f74099768d5bcec416d35389e01c82115be5e09" gracePeriod=2 Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.692707 4969 generic.go:334] "Generic (PLEG): container finished" podID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerID="f8c9bda401d4ff35c2e1d5075f74099768d5bcec416d35389e01c82115be5e09" exitCode=0 Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.692785 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfn5" event={"ID":"1aaa07d5-45f2-4074-af25-15ae361ff89e","Type":"ContainerDied","Data":"f8c9bda401d4ff35c2e1d5075f74099768d5bcec416d35389e01c82115be5e09"} Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.692829 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfn5" event={"ID":"1aaa07d5-45f2-4074-af25-15ae361ff89e","Type":"ContainerDied","Data":"65749c01f5e044b49e4f6693e144173f822b46e1ed9670fbf14f845db87474c5"} Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.692840 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65749c01f5e044b49e4f6693e144173f822b46e1ed9670fbf14f845db87474c5" Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.694596 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.694649 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvjn" event={"ID":"2333a480-04ca-4f93-937f-c91bf30323c2","Type":"ContainerStarted","Data":"634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08"} Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.795275 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96m55\" (UniqueName: \"kubernetes.io/projected/1aaa07d5-45f2-4074-af25-15ae361ff89e-kube-api-access-96m55\") pod \"1aaa07d5-45f2-4074-af25-15ae361ff89e\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.795319 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-utilities\") pod \"1aaa07d5-45f2-4074-af25-15ae361ff89e\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.795348 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-catalog-content\") pod \"1aaa07d5-45f2-4074-af25-15ae361ff89e\" (UID: \"1aaa07d5-45f2-4074-af25-15ae361ff89e\") " Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.796794 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-utilities" (OuterVolumeSpecName: "utilities") pod "1aaa07d5-45f2-4074-af25-15ae361ff89e" (UID: "1aaa07d5-45f2-4074-af25-15ae361ff89e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.808740 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aaa07d5-45f2-4074-af25-15ae361ff89e-kube-api-access-96m55" (OuterVolumeSpecName: "kube-api-access-96m55") pod "1aaa07d5-45f2-4074-af25-15ae361ff89e" (UID: "1aaa07d5-45f2-4074-af25-15ae361ff89e"). InnerVolumeSpecName "kube-api-access-96m55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.817508 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1aaa07d5-45f2-4074-af25-15ae361ff89e" (UID: "1aaa07d5-45f2-4074-af25-15ae361ff89e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.896784 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.896840 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96m55\" (UniqueName: \"kubernetes.io/projected/1aaa07d5-45f2-4074-af25-15ae361ff89e-kube-api-access-96m55\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:21 crc kubenswrapper[4969]: I1004 08:30:21.896866 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aaa07d5-45f2-4074-af25-15ae361ff89e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:22 crc kubenswrapper[4969]: I1004 08:30:22.703998 4969 generic.go:334] "Generic (PLEG): container finished" podID="2333a480-04ca-4f93-937f-c91bf30323c2" containerID="634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08" exitCode=0 Oct 04 08:30:22 crc kubenswrapper[4969]: I1004 08:30:22.704131 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlfn5" Oct 04 08:30:22 crc kubenswrapper[4969]: I1004 08:30:22.704258 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvjn" event={"ID":"2333a480-04ca-4f93-937f-c91bf30323c2","Type":"ContainerDied","Data":"634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08"} Oct 04 08:30:22 crc kubenswrapper[4969]: I1004 08:30:22.756922 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfn5"] Oct 04 08:30:22 crc kubenswrapper[4969]: I1004 08:30:22.766551 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfn5"] Oct 04 08:30:22 crc kubenswrapper[4969]: I1004 08:30:22.794861 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-74c8fbc96b-hr67n" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.087283 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aaa07d5-45f2-4074-af25-15ae361ff89e" path="/var/lib/kubelet/pods/1aaa07d5-45f2-4074-af25-15ae361ff89e/volumes" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.493584 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc"] Oct 04 08:30:23 crc kubenswrapper[4969]: E1004 08:30:23.494004 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerName="registry-server" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.494015 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerName="registry-server" Oct 04 08:30:23 crc kubenswrapper[4969]: E1004 08:30:23.494035 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerName="extract-content" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.494041 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerName="extract-content" Oct 04 08:30:23 crc kubenswrapper[4969]: E1004 08:30:23.494048 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerName="extract-utilities" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.494057 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerName="extract-utilities" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.494159 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aaa07d5-45f2-4074-af25-15ae361ff89e" containerName="registry-server" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.494582 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.502750 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.502897 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-r758x" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.504477 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-wdwv9"] Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.507391 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.508205 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc"] Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.509753 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.510004 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.576808 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-wtswn"] Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.577930 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.580165 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.580318 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.580345 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-2ss5x" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.580471 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.593607 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-cfvrx"] Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.594904 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.597549 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.612788 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cfvrx"] Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.619733 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-frr-conf\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.619804 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-frr-sockets\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.619834 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98r8n\" (UniqueName: \"kubernetes.io/projected/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-kube-api-access-98r8n\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.620104 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-frr-startup\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.620169 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhth6\" (UniqueName: \"kubernetes.io/projected/897cdf44-f275-487d-85e6-9f47825f8b87-kube-api-access-bhth6\") pod \"frr-k8s-webhook-server-64bf5d555-pz8cc\" (UID: \"897cdf44-f275-487d-85e6-9f47825f8b87\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.620212 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/897cdf44-f275-487d-85e6-9f47825f8b87-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pz8cc\" (UID: \"897cdf44-f275-487d-85e6-9f47825f8b87\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.620244 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-metrics-certs\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.620323 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-reloader\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.620450 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-metrics\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.710640 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvjn" event={"ID":"2333a480-04ca-4f93-937f-c91bf30323c2","Type":"ContainerStarted","Data":"09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba"} Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.721920 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-metrics-certs\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.721961 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7f4070ca-20a1-4c43-af82-890ef08c6fc3-metallb-excludel2\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.721995 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-frr-conf\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722016 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b-cert\") pod \"controller-68d546b9d8-cfvrx\" (UID: \"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b\") " pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722042 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-frr-sockets\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722064 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98r8n\" (UniqueName: \"kubernetes.io/projected/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-kube-api-access-98r8n\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722096 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-memberlist\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722135 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbg6x\" (UniqueName: \"kubernetes.io/projected/7f4070ca-20a1-4c43-af82-890ef08c6fc3-kube-api-access-mbg6x\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722194 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lt7h\" (UniqueName: \"kubernetes.io/projected/b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b-kube-api-access-9lt7h\") pod \"controller-68d546b9d8-cfvrx\" (UID: \"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b\") " pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722226 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-frr-startup\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722256 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhth6\" (UniqueName: \"kubernetes.io/projected/897cdf44-f275-487d-85e6-9f47825f8b87-kube-api-access-bhth6\") pod \"frr-k8s-webhook-server-64bf5d555-pz8cc\" (UID: \"897cdf44-f275-487d-85e6-9f47825f8b87\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722295 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/897cdf44-f275-487d-85e6-9f47825f8b87-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pz8cc\" (UID: \"897cdf44-f275-487d-85e6-9f47825f8b87\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722323 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-metrics-certs\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722353 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-reloader\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722386 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b-metrics-certs\") pod \"controller-68d546b9d8-cfvrx\" (UID: \"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b\") " pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.722518 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-metrics\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.723186 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-metrics\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.723221 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-frr-conf\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.723522 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-frr-sockets\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.724349 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-reloader\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.724705 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-frr-startup\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.729284 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-metrics-certs\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.732225 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/897cdf44-f275-487d-85e6-9f47825f8b87-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pz8cc\" (UID: \"897cdf44-f275-487d-85e6-9f47825f8b87\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.734506 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tnvjn" podStartSLOduration=2.261943926 podStartE2EDuration="4.734487777s" podCreationTimestamp="2025-10-04 08:30:19 +0000 UTC" firstStartedPulling="2025-10-04 08:30:20.685311682 +0000 UTC m=+848.439580536" lastFinishedPulling="2025-10-04 08:30:23.157855573 +0000 UTC m=+850.912124387" observedRunningTime="2025-10-04 08:30:23.733244287 +0000 UTC m=+851.487513121" watchObservedRunningTime="2025-10-04 08:30:23.734487777 +0000 UTC m=+851.488756591" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.743292 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98r8n\" (UniqueName: \"kubernetes.io/projected/6d4150d2-810e-4ceb-86a0-59155ecb1b0d-kube-api-access-98r8n\") pod \"frr-k8s-wdwv9\" (UID: \"6d4150d2-810e-4ceb-86a0-59155ecb1b0d\") " pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.750126 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhth6\" (UniqueName: \"kubernetes.io/projected/897cdf44-f275-487d-85e6-9f47825f8b87-kube-api-access-bhth6\") pod \"frr-k8s-webhook-server-64bf5d555-pz8cc\" (UID: \"897cdf44-f275-487d-85e6-9f47825f8b87\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.823971 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b-metrics-certs\") pod \"controller-68d546b9d8-cfvrx\" (UID: \"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b\") " pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.824045 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-metrics-certs\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.824092 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7f4070ca-20a1-4c43-af82-890ef08c6fc3-metallb-excludel2\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.825064 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b-cert\") pod \"controller-68d546b9d8-cfvrx\" (UID: \"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b\") " pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.825000 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7f4070ca-20a1-4c43-af82-890ef08c6fc3-metallb-excludel2\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.825114 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-memberlist\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.825175 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbg6x\" (UniqueName: \"kubernetes.io/projected/7f4070ca-20a1-4c43-af82-890ef08c6fc3-kube-api-access-mbg6x\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.825224 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lt7h\" (UniqueName: \"kubernetes.io/projected/b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b-kube-api-access-9lt7h\") pod \"controller-68d546b9d8-cfvrx\" (UID: \"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b\") " pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: E1004 08:30:23.825462 4969 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 08:30:23 crc kubenswrapper[4969]: E1004 08:30:23.825528 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-memberlist podName:7f4070ca-20a1-4c43-af82-890ef08c6fc3 nodeName:}" failed. No retries permitted until 2025-10-04 08:30:24.325510813 +0000 UTC m=+852.079779637 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-memberlist") pod "speaker-wtswn" (UID: "7f4070ca-20a1-4c43-af82-890ef08c6fc3") : secret "metallb-memberlist" not found Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.826930 4969 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.827340 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b-metrics-certs\") pod \"controller-68d546b9d8-cfvrx\" (UID: \"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b\") " pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.831726 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-metrics-certs\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.833951 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.839196 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b-cert\") pod \"controller-68d546b9d8-cfvrx\" (UID: \"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b\") " pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.842519 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbg6x\" (UniqueName: \"kubernetes.io/projected/7f4070ca-20a1-4c43-af82-890ef08c6fc3-kube-api-access-mbg6x\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.843938 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.851506 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lt7h\" (UniqueName: \"kubernetes.io/projected/b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b-kube-api-access-9lt7h\") pod \"controller-68d546b9d8-cfvrx\" (UID: \"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b\") " pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:23 crc kubenswrapper[4969]: I1004 08:30:23.916048 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.084592 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc"] Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.332661 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-memberlist\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:24 crc kubenswrapper[4969]: E1004 08:30:24.332871 4969 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 08:30:24 crc kubenswrapper[4969]: E1004 08:30:24.333089 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-memberlist podName:7f4070ca-20a1-4c43-af82-890ef08c6fc3 nodeName:}" failed. No retries permitted until 2025-10-04 08:30:25.333071659 +0000 UTC m=+853.087340473 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-memberlist") pod "speaker-wtswn" (UID: "7f4070ca-20a1-4c43-af82-890ef08c6fc3") : secret "metallb-memberlist" not found Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.393087 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cfvrx"] Oct 04 08:30:24 crc kubenswrapper[4969]: W1004 08:30:24.394302 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2acffa8_6ecf_4c46_bcf8_d4a6852a7a5b.slice/crio-efc04dab4e40a5645e9fe4a6b80a7301ba25f2191daf3d0fd1c0efec65caebb3 WatchSource:0}: Error finding container efc04dab4e40a5645e9fe4a6b80a7301ba25f2191daf3d0fd1c0efec65caebb3: Status 404 returned error can't find the container with id efc04dab4e40a5645e9fe4a6b80a7301ba25f2191daf3d0fd1c0efec65caebb3 Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.718058 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cfvrx" event={"ID":"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b","Type":"ContainerStarted","Data":"64e6e18ca372a58e87e9d50ce204f0eeb522c4e12329b090b05ea335000aab07"} Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.719304 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cfvrx" event={"ID":"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b","Type":"ContainerStarted","Data":"dd2e4edceb26efa9fae512cdbee1486f9b1f5ea1707d963cbd036d4daafcfc78"} Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.719476 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cfvrx" event={"ID":"b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b","Type":"ContainerStarted","Data":"efc04dab4e40a5645e9fe4a6b80a7301ba25f2191daf3d0fd1c0efec65caebb3"} Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.719589 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.719684 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerStarted","Data":"4393ba3586ecdfd589934a77857b27d0c65c7b4203881fd36967e6114c1b0ee3"} Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.720505 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" event={"ID":"897cdf44-f275-487d-85e6-9f47825f8b87","Type":"ContainerStarted","Data":"5dd412f771d72e6bdc835faec7ab4e8db5a8a616624a4538cddf16e09f38f9b4"} Oct 04 08:30:24 crc kubenswrapper[4969]: I1004 08:30:24.740027 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-cfvrx" podStartSLOduration=1.74000682 podStartE2EDuration="1.74000682s" podCreationTimestamp="2025-10-04 08:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:30:24.737872466 +0000 UTC m=+852.492141290" watchObservedRunningTime="2025-10-04 08:30:24.74000682 +0000 UTC m=+852.494275634" Oct 04 08:30:25 crc kubenswrapper[4969]: I1004 08:30:25.344930 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-memberlist\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:25 crc kubenswrapper[4969]: I1004 08:30:25.350851 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f4070ca-20a1-4c43-af82-890ef08c6fc3-memberlist\") pod \"speaker-wtswn\" (UID: \"7f4070ca-20a1-4c43-af82-890ef08c6fc3\") " pod="metallb-system/speaker-wtswn" Oct 04 08:30:25 crc kubenswrapper[4969]: I1004 08:30:25.396348 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wtswn" Oct 04 08:30:25 crc kubenswrapper[4969]: W1004 08:30:25.426977 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f4070ca_20a1_4c43_af82_890ef08c6fc3.slice/crio-a0aa0bf62e787cd804a4a542ac47988afb38250bc13d328fce5b5c60aee265c1 WatchSource:0}: Error finding container a0aa0bf62e787cd804a4a542ac47988afb38250bc13d328fce5b5c60aee265c1: Status 404 returned error can't find the container with id a0aa0bf62e787cd804a4a542ac47988afb38250bc13d328fce5b5c60aee265c1 Oct 04 08:30:25 crc kubenswrapper[4969]: I1004 08:30:25.732482 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wtswn" event={"ID":"7f4070ca-20a1-4c43-af82-890ef08c6fc3","Type":"ContainerStarted","Data":"c24e4d7ca5d042027c6c83c6ba4dd18105c20366cadb45fa39f61aae1d80b6d2"} Oct 04 08:30:25 crc kubenswrapper[4969]: I1004 08:30:25.732542 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wtswn" event={"ID":"7f4070ca-20a1-4c43-af82-890ef08c6fc3","Type":"ContainerStarted","Data":"a0aa0bf62e787cd804a4a542ac47988afb38250bc13d328fce5b5c60aee265c1"} Oct 04 08:30:26 crc kubenswrapper[4969]: I1004 08:30:26.748007 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wtswn" event={"ID":"7f4070ca-20a1-4c43-af82-890ef08c6fc3","Type":"ContainerStarted","Data":"c1e075fa8a332e4e53c718055b5640c97da3881926a9cb38a4f68c7b67a76818"} Oct 04 08:30:26 crc kubenswrapper[4969]: I1004 08:30:26.780689 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-wtswn" podStartSLOduration=3.7806748199999998 podStartE2EDuration="3.78067482s" podCreationTimestamp="2025-10-04 08:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:30:26.779117011 +0000 UTC m=+854.533385825" watchObservedRunningTime="2025-10-04 08:30:26.78067482 +0000 UTC m=+854.534943624" Oct 04 08:30:27 crc kubenswrapper[4969]: I1004 08:30:27.754001 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-wtswn" Oct 04 08:30:29 crc kubenswrapper[4969]: I1004 08:30:29.439892 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:29 crc kubenswrapper[4969]: I1004 08:30:29.442709 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:29 crc kubenswrapper[4969]: I1004 08:30:29.511025 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:29 crc kubenswrapper[4969]: I1004 08:30:29.813509 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:31 crc kubenswrapper[4969]: I1004 08:30:31.895093 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tnvjn"] Oct 04 08:30:32 crc kubenswrapper[4969]: I1004 08:30:32.795168 4969 generic.go:334] "Generic (PLEG): container finished" podID="6d4150d2-810e-4ceb-86a0-59155ecb1b0d" containerID="88821cc1edde45d05068b6203ccca6871e2b8e2c923ce6486a10a59f8f4f8cad" exitCode=0 Oct 04 08:30:32 crc kubenswrapper[4969]: I1004 08:30:32.795314 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerDied","Data":"88821cc1edde45d05068b6203ccca6871e2b8e2c923ce6486a10a59f8f4f8cad"} Oct 04 08:30:32 crc kubenswrapper[4969]: I1004 08:30:32.797413 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" event={"ID":"897cdf44-f275-487d-85e6-9f47825f8b87","Type":"ContainerStarted","Data":"c1ba93fa7663ae4c44b7768b4790590facf82ccbbdb1325dbf408b4fee8bee86"} Oct 04 08:30:32 crc kubenswrapper[4969]: I1004 08:30:32.797505 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tnvjn" podUID="2333a480-04ca-4f93-937f-c91bf30323c2" containerName="registry-server" containerID="cri-o://09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba" gracePeriod=2 Oct 04 08:30:32 crc kubenswrapper[4969]: I1004 08:30:32.797807 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:32 crc kubenswrapper[4969]: I1004 08:30:32.856211 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" podStartSLOduration=2.237072693 podStartE2EDuration="9.856192825s" podCreationTimestamp="2025-10-04 08:30:23 +0000 UTC" firstStartedPulling="2025-10-04 08:30:24.101050183 +0000 UTC m=+851.855318997" lastFinishedPulling="2025-10-04 08:30:31.720170315 +0000 UTC m=+859.474439129" observedRunningTime="2025-10-04 08:30:32.854907143 +0000 UTC m=+860.609176007" watchObservedRunningTime="2025-10-04 08:30:32.856192825 +0000 UTC m=+860.610461639" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.215545 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.269681 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvlb7\" (UniqueName: \"kubernetes.io/projected/2333a480-04ca-4f93-937f-c91bf30323c2-kube-api-access-qvlb7\") pod \"2333a480-04ca-4f93-937f-c91bf30323c2\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.269807 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-catalog-content\") pod \"2333a480-04ca-4f93-937f-c91bf30323c2\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.269875 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-utilities\") pod \"2333a480-04ca-4f93-937f-c91bf30323c2\" (UID: \"2333a480-04ca-4f93-937f-c91bf30323c2\") " Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.271080 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-utilities" (OuterVolumeSpecName: "utilities") pod "2333a480-04ca-4f93-937f-c91bf30323c2" (UID: "2333a480-04ca-4f93-937f-c91bf30323c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.276795 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2333a480-04ca-4f93-937f-c91bf30323c2-kube-api-access-qvlb7" (OuterVolumeSpecName: "kube-api-access-qvlb7") pod "2333a480-04ca-4f93-937f-c91bf30323c2" (UID: "2333a480-04ca-4f93-937f-c91bf30323c2"). InnerVolumeSpecName "kube-api-access-qvlb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.317763 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2333a480-04ca-4f93-937f-c91bf30323c2" (UID: "2333a480-04ca-4f93-937f-c91bf30323c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.371435 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvlb7\" (UniqueName: \"kubernetes.io/projected/2333a480-04ca-4f93-937f-c91bf30323c2-kube-api-access-qvlb7\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.371655 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.371665 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2333a480-04ca-4f93-937f-c91bf30323c2-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.818215 4969 generic.go:334] "Generic (PLEG): container finished" podID="2333a480-04ca-4f93-937f-c91bf30323c2" containerID="09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba" exitCode=0 Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.818291 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvjn" event={"ID":"2333a480-04ca-4f93-937f-c91bf30323c2","Type":"ContainerDied","Data":"09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba"} Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.818321 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvjn" event={"ID":"2333a480-04ca-4f93-937f-c91bf30323c2","Type":"ContainerDied","Data":"c3f8a50ebbcf8240f86b3d87b1a1403c4d42761f76026466b36c9244aaf63da3"} Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.818343 4969 scope.go:117] "RemoveContainer" containerID="09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.818520 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvjn" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.841934 4969 scope.go:117] "RemoveContainer" containerID="634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.842503 4969 generic.go:334] "Generic (PLEG): container finished" podID="6d4150d2-810e-4ceb-86a0-59155ecb1b0d" containerID="a4b3c0c27ec005e54664bdaea8c21e0102a946fa4ba421510e31c685e0ef128c" exitCode=0 Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.848617 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerDied","Data":"a4b3c0c27ec005e54664bdaea8c21e0102a946fa4ba421510e31c685e0ef128c"} Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.896613 4969 scope.go:117] "RemoveContainer" containerID="3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.936565 4969 scope.go:117] "RemoveContainer" containerID="09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba" Oct 04 08:30:33 crc kubenswrapper[4969]: E1004 08:30:33.958485 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba\": container with ID starting with 09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba not found: ID does not exist" containerID="09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.958751 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba"} err="failed to get container status \"09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba\": rpc error: code = NotFound desc = could not find container \"09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba\": container with ID starting with 09bee8aedf8e73ad131f63ec625cc667d87e224111986a9282eb2e5a9f0965ba not found: ID does not exist" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.958834 4969 scope.go:117] "RemoveContainer" containerID="634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.958694 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tnvjn"] Oct 04 08:30:33 crc kubenswrapper[4969]: E1004 08:30:33.963549 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08\": container with ID starting with 634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08 not found: ID does not exist" containerID="634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.963591 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08"} err="failed to get container status \"634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08\": rpc error: code = NotFound desc = could not find container \"634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08\": container with ID starting with 634bc19cfce771351fd4ff3bc798297b5b8d4d3a997ad82ac135b924e7b45f08 not found: ID does not exist" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.963616 4969 scope.go:117] "RemoveContainer" containerID="3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f" Oct 04 08:30:33 crc kubenswrapper[4969]: E1004 08:30:33.964294 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f\": container with ID starting with 3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f not found: ID does not exist" containerID="3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.964323 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f"} err="failed to get container status \"3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f\": rpc error: code = NotFound desc = could not find container \"3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f\": container with ID starting with 3b9a4c81058d3d70c248d4bfbec158229435e4d8606f8009a89525a17f7dec0f not found: ID does not exist" Oct 04 08:30:33 crc kubenswrapper[4969]: I1004 08:30:33.965189 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tnvjn"] Oct 04 08:30:34 crc kubenswrapper[4969]: I1004 08:30:34.856311 4969 generic.go:334] "Generic (PLEG): container finished" podID="6d4150d2-810e-4ceb-86a0-59155ecb1b0d" containerID="22c1beefa513cde65ccc764c53b141b0d8792c547e086bd1a57ec59b7faddbb8" exitCode=0 Oct 04 08:30:34 crc kubenswrapper[4969]: I1004 08:30:34.856381 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerDied","Data":"22c1beefa513cde65ccc764c53b141b0d8792c547e086bd1a57ec59b7faddbb8"} Oct 04 08:30:35 crc kubenswrapper[4969]: I1004 08:30:35.068520 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2333a480-04ca-4f93-937f-c91bf30323c2" path="/var/lib/kubelet/pods/2333a480-04ca-4f93-937f-c91bf30323c2/volumes" Oct 04 08:30:35 crc kubenswrapper[4969]: I1004 08:30:35.405302 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-wtswn" Oct 04 08:30:35 crc kubenswrapper[4969]: I1004 08:30:35.870168 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerStarted","Data":"3ce8cb9921a1170a950a28d828a38e2bf4d8ea24803ac47a736c82cba6acba5f"} Oct 04 08:30:35 crc kubenswrapper[4969]: I1004 08:30:35.870556 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerStarted","Data":"e2e5268aedf1b094ce86340a4b194bac6745aaa2405004544a2b357f53da3edb"} Oct 04 08:30:35 crc kubenswrapper[4969]: I1004 08:30:35.870574 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerStarted","Data":"8a62d544b0f904d29116190863074713fbbcba9a2a48fe4fb4b19490739d9d2b"} Oct 04 08:30:35 crc kubenswrapper[4969]: I1004 08:30:35.870588 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerStarted","Data":"43292d8477b192696b803f3806fff63be176d9979de71e13dc0d807383bcaaaa"} Oct 04 08:30:35 crc kubenswrapper[4969]: I1004 08:30:35.870603 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerStarted","Data":"38d38b7c8362e56bd45449f274bad7323f47a8b7b11227730dc88673b067b14c"} Oct 04 08:30:36 crc kubenswrapper[4969]: I1004 08:30:36.881945 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wdwv9" event={"ID":"6d4150d2-810e-4ceb-86a0-59155ecb1b0d","Type":"ContainerStarted","Data":"6f77d4fd968d1aad8383ad5cce0dfdb88f6d13a79229afb7a0798460695923fa"} Oct 04 08:30:36 crc kubenswrapper[4969]: I1004 08:30:36.882129 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:36 crc kubenswrapper[4969]: I1004 08:30:36.954701 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-wdwv9" podStartSLOduration=6.209739659 podStartE2EDuration="13.954680363s" podCreationTimestamp="2025-10-04 08:30:23 +0000 UTC" firstStartedPulling="2025-10-04 08:30:23.971015626 +0000 UTC m=+851.725284440" lastFinishedPulling="2025-10-04 08:30:31.71595633 +0000 UTC m=+859.470225144" observedRunningTime="2025-10-04 08:30:36.942600662 +0000 UTC m=+864.696869476" watchObservedRunningTime="2025-10-04 08:30:36.954680363 +0000 UTC m=+864.708949177" Oct 04 08:30:38 crc kubenswrapper[4969]: I1004 08:30:38.835186 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:38 crc kubenswrapper[4969]: I1004 08:30:38.890974 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.512077 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-d748d"] Oct 04 08:30:42 crc kubenswrapper[4969]: E1004 08:30:42.512987 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2333a480-04ca-4f93-937f-c91bf30323c2" containerName="extract-utilities" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.513022 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="2333a480-04ca-4f93-937f-c91bf30323c2" containerName="extract-utilities" Oct 04 08:30:42 crc kubenswrapper[4969]: E1004 08:30:42.513061 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2333a480-04ca-4f93-937f-c91bf30323c2" containerName="extract-content" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.513079 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="2333a480-04ca-4f93-937f-c91bf30323c2" containerName="extract-content" Oct 04 08:30:42 crc kubenswrapper[4969]: E1004 08:30:42.513130 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2333a480-04ca-4f93-937f-c91bf30323c2" containerName="registry-server" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.513148 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="2333a480-04ca-4f93-937f-c91bf30323c2" containerName="registry-server" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.513412 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="2333a480-04ca-4f93-937f-c91bf30323c2" containerName="registry-server" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.514319 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-d748d" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.517615 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.518053 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.522184 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-7znlq" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.527909 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-d748d"] Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.641960 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwkqw\" (UniqueName: \"kubernetes.io/projected/03968d39-1ce4-4b22-875f-5b6f391a21c3-kube-api-access-fwkqw\") pod \"openstack-operator-index-d748d\" (UID: \"03968d39-1ce4-4b22-875f-5b6f391a21c3\") " pod="openstack-operators/openstack-operator-index-d748d" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.743160 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwkqw\" (UniqueName: \"kubernetes.io/projected/03968d39-1ce4-4b22-875f-5b6f391a21c3-kube-api-access-fwkqw\") pod \"openstack-operator-index-d748d\" (UID: \"03968d39-1ce4-4b22-875f-5b6f391a21c3\") " pod="openstack-operators/openstack-operator-index-d748d" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.776202 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwkqw\" (UniqueName: \"kubernetes.io/projected/03968d39-1ce4-4b22-875f-5b6f391a21c3-kube-api-access-fwkqw\") pod \"openstack-operator-index-d748d\" (UID: \"03968d39-1ce4-4b22-875f-5b6f391a21c3\") " pod="openstack-operators/openstack-operator-index-d748d" Oct 04 08:30:42 crc kubenswrapper[4969]: I1004 08:30:42.858465 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-d748d" Oct 04 08:30:43 crc kubenswrapper[4969]: I1004 08:30:43.365237 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-d748d"] Oct 04 08:30:43 crc kubenswrapper[4969]: W1004 08:30:43.374736 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03968d39_1ce4_4b22_875f_5b6f391a21c3.slice/crio-fbec35fcc89c74ae3fa7963d9d7821028028715b2818de1744a6d78dc79eb0cc WatchSource:0}: Error finding container fbec35fcc89c74ae3fa7963d9d7821028028715b2818de1744a6d78dc79eb0cc: Status 404 returned error can't find the container with id fbec35fcc89c74ae3fa7963d9d7821028028715b2818de1744a6d78dc79eb0cc Oct 04 08:30:43 crc kubenswrapper[4969]: I1004 08:30:43.859915 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pz8cc" Oct 04 08:30:43 crc kubenswrapper[4969]: I1004 08:30:43.930011 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-cfvrx" Oct 04 08:30:43 crc kubenswrapper[4969]: I1004 08:30:43.952569 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-d748d" event={"ID":"03968d39-1ce4-4b22-875f-5b6f391a21c3","Type":"ContainerStarted","Data":"fbec35fcc89c74ae3fa7963d9d7821028028715b2818de1744a6d78dc79eb0cc"} Oct 04 08:30:46 crc kubenswrapper[4969]: I1004 08:30:46.974212 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-d748d" event={"ID":"03968d39-1ce4-4b22-875f-5b6f391a21c3","Type":"ContainerStarted","Data":"5eafbca30b8f99e66eea54c520c705c0e136e775afa6305b5700060ff2990d30"} Oct 04 08:30:47 crc kubenswrapper[4969]: I1004 08:30:47.008085 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-d748d" podStartSLOduration=2.539613683 podStartE2EDuration="5.008050613s" podCreationTimestamp="2025-10-04 08:30:42 +0000 UTC" firstStartedPulling="2025-10-04 08:30:43.378463827 +0000 UTC m=+871.132732671" lastFinishedPulling="2025-10-04 08:30:45.846900787 +0000 UTC m=+873.601169601" observedRunningTime="2025-10-04 08:30:46.996581548 +0000 UTC m=+874.750850422" watchObservedRunningTime="2025-10-04 08:30:47.008050613 +0000 UTC m=+874.762319477" Oct 04 08:30:52 crc kubenswrapper[4969]: I1004 08:30:52.859691 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-d748d" Oct 04 08:30:52 crc kubenswrapper[4969]: I1004 08:30:52.860513 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-d748d" Oct 04 08:30:52 crc kubenswrapper[4969]: I1004 08:30:52.901361 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-d748d" Oct 04 08:30:53 crc kubenswrapper[4969]: I1004 08:30:53.069023 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-d748d" Oct 04 08:30:53 crc kubenswrapper[4969]: I1004 08:30:53.838995 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wdwv9" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.331475 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng"] Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.332942 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.336105 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-tr24d" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.362737 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng"] Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.446178 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-util\") pod \"d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.446245 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l4rd\" (UniqueName: \"kubernetes.io/projected/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-kube-api-access-8l4rd\") pod \"d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.446342 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-bundle\") pod \"d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.548142 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-util\") pod \"d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.548326 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l4rd\" (UniqueName: \"kubernetes.io/projected/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-kube-api-access-8l4rd\") pod \"d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.548589 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-bundle\") pod \"d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.549115 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-bundle\") pod \"d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.549112 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-util\") pod \"d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.602964 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l4rd\" (UniqueName: \"kubernetes.io/projected/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-kube-api-access-8l4rd\") pod \"d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:54 crc kubenswrapper[4969]: I1004 08:30:54.651611 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:55 crc kubenswrapper[4969]: I1004 08:30:55.110443 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng"] Oct 04 08:30:55 crc kubenswrapper[4969]: W1004 08:30:55.111544 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded5c8e4a_dc27_4aa1_a533_9f918932ddff.slice/crio-6a86f9aa9830ae97c048633c2978ed59adff4686c0306d8ed71b3b7f590f4ae2 WatchSource:0}: Error finding container 6a86f9aa9830ae97c048633c2978ed59adff4686c0306d8ed71b3b7f590f4ae2: Status 404 returned error can't find the container with id 6a86f9aa9830ae97c048633c2978ed59adff4686c0306d8ed71b3b7f590f4ae2 Oct 04 08:30:56 crc kubenswrapper[4969]: I1004 08:30:56.051841 4969 generic.go:334] "Generic (PLEG): container finished" podID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerID="86ffc0b5a07af2832199ba92b7abc574ac4b0dbc9766aee8e3136b9ecc3e52de" exitCode=0 Oct 04 08:30:56 crc kubenswrapper[4969]: I1004 08:30:56.051990 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" event={"ID":"ed5c8e4a-dc27-4aa1-a533-9f918932ddff","Type":"ContainerDied","Data":"86ffc0b5a07af2832199ba92b7abc574ac4b0dbc9766aee8e3136b9ecc3e52de"} Oct 04 08:30:56 crc kubenswrapper[4969]: I1004 08:30:56.052070 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" event={"ID":"ed5c8e4a-dc27-4aa1-a533-9f918932ddff","Type":"ContainerStarted","Data":"6a86f9aa9830ae97c048633c2978ed59adff4686c0306d8ed71b3b7f590f4ae2"} Oct 04 08:30:57 crc kubenswrapper[4969]: I1004 08:30:57.067066 4969 generic.go:334] "Generic (PLEG): container finished" podID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerID="f919ceceaa1b9717bb38a424f6623fd04f4fcb847ec42b8fd6eb75302aca1fa9" exitCode=0 Oct 04 08:30:57 crc kubenswrapper[4969]: I1004 08:30:57.069461 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" event={"ID":"ed5c8e4a-dc27-4aa1-a533-9f918932ddff","Type":"ContainerDied","Data":"f919ceceaa1b9717bb38a424f6623fd04f4fcb847ec42b8fd6eb75302aca1fa9"} Oct 04 08:30:58 crc kubenswrapper[4969]: I1004 08:30:58.077891 4969 generic.go:334] "Generic (PLEG): container finished" podID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerID="6fc63c6f4a71100a2e6494c576d53e505d5ba5f8c5d656bc888869955bf46354" exitCode=0 Oct 04 08:30:58 crc kubenswrapper[4969]: I1004 08:30:58.077952 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" event={"ID":"ed5c8e4a-dc27-4aa1-a533-9f918932ddff","Type":"ContainerDied","Data":"6fc63c6f4a71100a2e6494c576d53e505d5ba5f8c5d656bc888869955bf46354"} Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.434870 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.542271 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-bundle\") pod \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.542513 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l4rd\" (UniqueName: \"kubernetes.io/projected/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-kube-api-access-8l4rd\") pod \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.543174 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-util\") pod \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\" (UID: \"ed5c8e4a-dc27-4aa1-a533-9f918932ddff\") " Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.543552 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-bundle" (OuterVolumeSpecName: "bundle") pod "ed5c8e4a-dc27-4aa1-a533-9f918932ddff" (UID: "ed5c8e4a-dc27-4aa1-a533-9f918932ddff"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.544260 4969 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.551343 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-kube-api-access-8l4rd" (OuterVolumeSpecName: "kube-api-access-8l4rd") pod "ed5c8e4a-dc27-4aa1-a533-9f918932ddff" (UID: "ed5c8e4a-dc27-4aa1-a533-9f918932ddff"). InnerVolumeSpecName "kube-api-access-8l4rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.566616 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-util" (OuterVolumeSpecName: "util") pod "ed5c8e4a-dc27-4aa1-a533-9f918932ddff" (UID: "ed5c8e4a-dc27-4aa1-a533-9f918932ddff"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.645688 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l4rd\" (UniqueName: \"kubernetes.io/projected/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-kube-api-access-8l4rd\") on node \"crc\" DevicePath \"\"" Oct 04 08:30:59 crc kubenswrapper[4969]: I1004 08:30:59.645742 4969 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed5c8e4a-dc27-4aa1-a533-9f918932ddff-util\") on node \"crc\" DevicePath \"\"" Oct 04 08:31:00 crc kubenswrapper[4969]: I1004 08:31:00.098903 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" event={"ID":"ed5c8e4a-dc27-4aa1-a533-9f918932ddff","Type":"ContainerDied","Data":"6a86f9aa9830ae97c048633c2978ed59adff4686c0306d8ed71b3b7f590f4ae2"} Oct 04 08:31:00 crc kubenswrapper[4969]: I1004 08:31:00.098966 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a86f9aa9830ae97c048633c2978ed59adff4686c0306d8ed71b3b7f590f4ae2" Oct 04 08:31:00 crc kubenswrapper[4969]: I1004 08:31:00.098971 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.623289 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89"] Oct 04 08:31:02 crc kubenswrapper[4969]: E1004 08:31:02.624039 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerName="pull" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.624053 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerName="pull" Oct 04 08:31:02 crc kubenswrapper[4969]: E1004 08:31:02.624069 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerName="extract" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.624077 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerName="extract" Oct 04 08:31:02 crc kubenswrapper[4969]: E1004 08:31:02.624100 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerName="util" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.624109 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerName="util" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.624262 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed5c8e4a-dc27-4aa1-a533-9f918932ddff" containerName="extract" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.625205 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.627282 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-xv9qn" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.647884 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89"] Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.788777 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t77v4\" (UniqueName: \"kubernetes.io/projected/bccc7a17-2a0f-4fbb-80da-a1775673e40f-kube-api-access-t77v4\") pod \"openstack-operator-controller-operator-76fbc76964-zng89\" (UID: \"bccc7a17-2a0f-4fbb-80da-a1775673e40f\") " pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.890189 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t77v4\" (UniqueName: \"kubernetes.io/projected/bccc7a17-2a0f-4fbb-80da-a1775673e40f-kube-api-access-t77v4\") pod \"openstack-operator-controller-operator-76fbc76964-zng89\" (UID: \"bccc7a17-2a0f-4fbb-80da-a1775673e40f\") " pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.910799 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t77v4\" (UniqueName: \"kubernetes.io/projected/bccc7a17-2a0f-4fbb-80da-a1775673e40f-kube-api-access-t77v4\") pod \"openstack-operator-controller-operator-76fbc76964-zng89\" (UID: \"bccc7a17-2a0f-4fbb-80da-a1775673e40f\") " pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" Oct 04 08:31:02 crc kubenswrapper[4969]: I1004 08:31:02.944470 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" Oct 04 08:31:03 crc kubenswrapper[4969]: I1004 08:31:03.373573 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89"] Oct 04 08:31:04 crc kubenswrapper[4969]: I1004 08:31:04.134527 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" event={"ID":"bccc7a17-2a0f-4fbb-80da-a1775673e40f","Type":"ContainerStarted","Data":"906c91ac1347dfcb794c80837a24733bc6c5879159cebe412457cbda2aff013b"} Oct 04 08:31:07 crc kubenswrapper[4969]: I1004 08:31:07.155437 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" event={"ID":"bccc7a17-2a0f-4fbb-80da-a1775673e40f","Type":"ContainerStarted","Data":"d553e3ab73e5553d2438961a86beb115b2c426a4f1144a00e10bf9ccc95fe57a"} Oct 04 08:31:09 crc kubenswrapper[4969]: I1004 08:31:09.171318 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" event={"ID":"bccc7a17-2a0f-4fbb-80da-a1775673e40f","Type":"ContainerStarted","Data":"6adab2bc9ef78bbe267870f0e02d67b50b029fb6103434a618ce521e3653382d"} Oct 04 08:31:09 crc kubenswrapper[4969]: I1004 08:31:09.171791 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" Oct 04 08:31:09 crc kubenswrapper[4969]: I1004 08:31:09.201226 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" podStartSLOduration=1.577596418 podStartE2EDuration="7.201210342s" podCreationTimestamp="2025-10-04 08:31:02 +0000 UTC" firstStartedPulling="2025-10-04 08:31:03.374521012 +0000 UTC m=+891.128789826" lastFinishedPulling="2025-10-04 08:31:08.998134896 +0000 UTC m=+896.752403750" observedRunningTime="2025-10-04 08:31:09.197970871 +0000 UTC m=+896.952239685" watchObservedRunningTime="2025-10-04 08:31:09.201210342 +0000 UTC m=+896.955479156" Oct 04 08:31:12 crc kubenswrapper[4969]: I1004 08:31:12.948411 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-76fbc76964-zng89" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.023869 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.026136 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.027139 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.028104 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.029402 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hwrfb" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.032455 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-xzd5l" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.033668 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.043782 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.063266 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.064157 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.068397 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-llzbs" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.085846 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.086863 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.090996 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-crnl2" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.095610 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.102143 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.122835 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.123744 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.126758 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-h8ljj" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.128451 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.135067 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.136281 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.144954 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-s4fdr" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.149944 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.150877 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.160435 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vnjnx" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.160616 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.185532 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.190353 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkgk7\" (UniqueName: \"kubernetes.io/projected/eb8a54aa-2ce0-400c-a06f-cc3513b1df30-kube-api-access-tkgk7\") pod \"barbican-operator-controller-manager-5f7c849b98-x7vqg\" (UID: \"eb8a54aa-2ce0-400c-a06f-cc3513b1df30\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.190487 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxkgn\" (UniqueName: \"kubernetes.io/projected/d7dc7acf-3666-46fd-99cd-8447344cd10c-kube-api-access-kxkgn\") pod \"cinder-operator-controller-manager-7d4d4f8d-d7885\" (UID: \"d7dc7acf-3666-46fd-99cd-8447344cd10c\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.190527 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65t6c\" (UniqueName: \"kubernetes.io/projected/b7837ab8-145d-4f08-b407-5f0325119fb1-kube-api-access-65t6c\") pod \"designate-operator-controller-manager-75dfd9b554-m6llv\" (UID: \"b7837ab8-145d-4f08-b407-5f0325119fb1\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.190554 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntd9q\" (UniqueName: \"kubernetes.io/projected/0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f-kube-api-access-ntd9q\") pod \"glance-operator-controller-manager-5568b5d68-lvh4x\" (UID: \"0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.197804 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.201547 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.202565 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.207730 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-wxldp" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.208497 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.222902 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.223823 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.230839 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-khj7c" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.238073 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.239060 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.245719 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-qfljl" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.258335 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.262466 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.276535 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.277568 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.281858 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-r758w" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.292227 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89f37d97-0acc-4940-bd3a-733f4e6ed592-cert\") pod \"infra-operator-controller-manager-658588b8c9-zg6hn\" (UID: \"89f37d97-0acc-4940-bd3a-733f4e6ed592\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.292264 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlfzs\" (UniqueName: \"kubernetes.io/projected/18413b49-54b1-4ac9-8225-4aa748d0e4f3-kube-api-access-mlfzs\") pod \"horizon-operator-controller-manager-54876c876f-m8tn8\" (UID: \"18413b49-54b1-4ac9-8225-4aa748d0e4f3\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.292298 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxkgn\" (UniqueName: \"kubernetes.io/projected/d7dc7acf-3666-46fd-99cd-8447344cd10c-kube-api-access-kxkgn\") pod \"cinder-operator-controller-manager-7d4d4f8d-d7885\" (UID: \"d7dc7acf-3666-46fd-99cd-8447344cd10c\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.292319 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65t6c\" (UniqueName: \"kubernetes.io/projected/b7837ab8-145d-4f08-b407-5f0325119fb1-kube-api-access-65t6c\") pod \"designate-operator-controller-manager-75dfd9b554-m6llv\" (UID: \"b7837ab8-145d-4f08-b407-5f0325119fb1\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.292632 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntd9q\" (UniqueName: \"kubernetes.io/projected/0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f-kube-api-access-ntd9q\") pod \"glance-operator-controller-manager-5568b5d68-lvh4x\" (UID: \"0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.296154 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chnnk\" (UniqueName: \"kubernetes.io/projected/89f37d97-0acc-4940-bd3a-733f4e6ed592-kube-api-access-chnnk\") pod \"infra-operator-controller-manager-658588b8c9-zg6hn\" (UID: \"89f37d97-0acc-4940-bd3a-733f4e6ed592\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.296189 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6wgb\" (UniqueName: \"kubernetes.io/projected/fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea-kube-api-access-s6wgb\") pod \"heat-operator-controller-manager-8f58bc9db-htxsn\" (UID: \"fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.296289 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkgk7\" (UniqueName: \"kubernetes.io/projected/eb8a54aa-2ce0-400c-a06f-cc3513b1df30-kube-api-access-tkgk7\") pod \"barbican-operator-controller-manager-5f7c849b98-x7vqg\" (UID: \"eb8a54aa-2ce0-400c-a06f-cc3513b1df30\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.296073 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.301503 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.302547 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.313547 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.319681 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-9r6vs" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.333676 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65t6c\" (UniqueName: \"kubernetes.io/projected/b7837ab8-145d-4f08-b407-5f0325119fb1-kube-api-access-65t6c\") pod \"designate-operator-controller-manager-75dfd9b554-m6llv\" (UID: \"b7837ab8-145d-4f08-b407-5f0325119fb1\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.334122 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkgk7\" (UniqueName: \"kubernetes.io/projected/eb8a54aa-2ce0-400c-a06f-cc3513b1df30-kube-api-access-tkgk7\") pod \"barbican-operator-controller-manager-5f7c849b98-x7vqg\" (UID: \"eb8a54aa-2ce0-400c-a06f-cc3513b1df30\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.334402 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.370273 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntd9q\" (UniqueName: \"kubernetes.io/projected/0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f-kube-api-access-ntd9q\") pod \"glance-operator-controller-manager-5568b5d68-lvh4x\" (UID: \"0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.383889 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.386107 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.401403 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr266\" (UniqueName: \"kubernetes.io/projected/0ce59178-f382-4436-be86-5bc4c5d4a9b0-kube-api-access-lr266\") pod \"neutron-operator-controller-manager-8d984cc4d-54zxr\" (UID: \"0ce59178-f382-4436-be86-5bc4c5d4a9b0\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.401467 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l5fs\" (UniqueName: \"kubernetes.io/projected/edfa72c9-187e-4d6c-8aab-66cb36073282-kube-api-access-9l5fs\") pod \"ironic-operator-controller-manager-699b87f775-x8xxq\" (UID: \"edfa72c9-187e-4d6c-8aab-66cb36073282\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.401509 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89f37d97-0acc-4940-bd3a-733f4e6ed592-cert\") pod \"infra-operator-controller-manager-658588b8c9-zg6hn\" (UID: \"89f37d97-0acc-4940-bd3a-733f4e6ed592\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.401534 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlfzs\" (UniqueName: \"kubernetes.io/projected/18413b49-54b1-4ac9-8225-4aa748d0e4f3-kube-api-access-mlfzs\") pod \"horizon-operator-controller-manager-54876c876f-m8tn8\" (UID: \"18413b49-54b1-4ac9-8225-4aa748d0e4f3\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.401555 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9xvp\" (UniqueName: \"kubernetes.io/projected/5603909a-7f2a-4c86-8446-9fae64c02482-kube-api-access-z9xvp\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t\" (UID: \"5603909a-7f2a-4c86-8446-9fae64c02482\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.401595 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4dzd\" (UniqueName: \"kubernetes.io/projected/5b816f82-7f2c-4537-aadd-53b6bded2e14-kube-api-access-c4dzd\") pod \"keystone-operator-controller-manager-655d88ccb9-t95db\" (UID: \"5b816f82-7f2c-4537-aadd-53b6bded2e14\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.401622 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chnnk\" (UniqueName: \"kubernetes.io/projected/89f37d97-0acc-4940-bd3a-733f4e6ed592-kube-api-access-chnnk\") pod \"infra-operator-controller-manager-658588b8c9-zg6hn\" (UID: \"89f37d97-0acc-4940-bd3a-733f4e6ed592\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.401644 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6wgb\" (UniqueName: \"kubernetes.io/projected/fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea-kube-api-access-s6wgb\") pod \"heat-operator-controller-manager-8f58bc9db-htxsn\" (UID: \"fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.401676 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs8rx\" (UniqueName: \"kubernetes.io/projected/d4376d0e-3428-49ec-993f-48b32d4c8863-kube-api-access-gs8rx\") pod \"manila-operator-controller-manager-65d89cfd9f-lqq2s\" (UID: \"d4376d0e-3428-49ec-993f-48b32d4c8863\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" Oct 04 08:31:30 crc kubenswrapper[4969]: E1004 08:31:30.401851 4969 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 04 08:31:30 crc kubenswrapper[4969]: E1004 08:31:30.401898 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89f37d97-0acc-4940-bd3a-733f4e6ed592-cert podName:89f37d97-0acc-4940-bd3a-733f4e6ed592 nodeName:}" failed. No retries permitted until 2025-10-04 08:31:30.901881904 +0000 UTC m=+918.656150718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/89f37d97-0acc-4940-bd3a-733f4e6ed592-cert") pod "infra-operator-controller-manager-658588b8c9-zg6hn" (UID: "89f37d97-0acc-4940-bd3a-733f4e6ed592") : secret "infra-operator-webhook-server-cert" not found Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.404170 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxkgn\" (UniqueName: \"kubernetes.io/projected/d7dc7acf-3666-46fd-99cd-8447344cd10c-kube-api-access-kxkgn\") pod \"cinder-operator-controller-manager-7d4d4f8d-d7885\" (UID: \"d7dc7acf-3666-46fd-99cd-8447344cd10c\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.410441 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.415801 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-bxxsc" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.430231 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.453099 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6wgb\" (UniqueName: \"kubernetes.io/projected/fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea-kube-api-access-s6wgb\") pod \"heat-operator-controller-manager-8f58bc9db-htxsn\" (UID: \"fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.453545 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chnnk\" (UniqueName: \"kubernetes.io/projected/89f37d97-0acc-4940-bd3a-733f4e6ed592-kube-api-access-chnnk\") pod \"infra-operator-controller-manager-658588b8c9-zg6hn\" (UID: \"89f37d97-0acc-4940-bd3a-733f4e6ed592\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.481172 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlfzs\" (UniqueName: \"kubernetes.io/projected/18413b49-54b1-4ac9-8225-4aa748d0e4f3-kube-api-access-mlfzs\") pod \"horizon-operator-controller-manager-54876c876f-m8tn8\" (UID: \"18413b49-54b1-4ac9-8225-4aa748d0e4f3\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.491832 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.492868 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.495458 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-csf2f" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.501106 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.505663 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs8rx\" (UniqueName: \"kubernetes.io/projected/d4376d0e-3428-49ec-993f-48b32d4c8863-kube-api-access-gs8rx\") pod \"manila-operator-controller-manager-65d89cfd9f-lqq2s\" (UID: \"d4376d0e-3428-49ec-993f-48b32d4c8863\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.505753 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr266\" (UniqueName: \"kubernetes.io/projected/0ce59178-f382-4436-be86-5bc4c5d4a9b0-kube-api-access-lr266\") pod \"neutron-operator-controller-manager-8d984cc4d-54zxr\" (UID: \"0ce59178-f382-4436-be86-5bc4c5d4a9b0\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.505787 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l5fs\" (UniqueName: \"kubernetes.io/projected/edfa72c9-187e-4d6c-8aab-66cb36073282-kube-api-access-9l5fs\") pod \"ironic-operator-controller-manager-699b87f775-x8xxq\" (UID: \"edfa72c9-187e-4d6c-8aab-66cb36073282\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.505820 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7bz7\" (UniqueName: \"kubernetes.io/projected/34735d37-aafa-4641-a951-f4d4d7fdab3b-kube-api-access-r7bz7\") pod \"nova-operator-controller-manager-7c7fc454ff-bg7zq\" (UID: \"34735d37-aafa-4641-a951-f4d4d7fdab3b\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.505879 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9xvp\" (UniqueName: \"kubernetes.io/projected/5603909a-7f2a-4c86-8446-9fae64c02482-kube-api-access-z9xvp\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t\" (UID: \"5603909a-7f2a-4c86-8446-9fae64c02482\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.505921 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4dzd\" (UniqueName: \"kubernetes.io/projected/5b816f82-7f2c-4537-aadd-53b6bded2e14-kube-api-access-c4dzd\") pod \"keystone-operator-controller-manager-655d88ccb9-t95db\" (UID: \"5b816f82-7f2c-4537-aadd-53b6bded2e14\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.518786 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.526089 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.527218 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.531662 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-r6scd" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.532189 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.536477 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l5fs\" (UniqueName: \"kubernetes.io/projected/edfa72c9-187e-4d6c-8aab-66cb36073282-kube-api-access-9l5fs\") pod \"ironic-operator-controller-manager-699b87f775-x8xxq\" (UID: \"edfa72c9-187e-4d6c-8aab-66cb36073282\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.540319 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr266\" (UniqueName: \"kubernetes.io/projected/0ce59178-f382-4436-be86-5bc4c5d4a9b0-kube-api-access-lr266\") pod \"neutron-operator-controller-manager-8d984cc4d-54zxr\" (UID: \"0ce59178-f382-4436-be86-5bc4c5d4a9b0\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.559880 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.561117 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.562019 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs8rx\" (UniqueName: \"kubernetes.io/projected/d4376d0e-3428-49ec-993f-48b32d4c8863-kube-api-access-gs8rx\") pod \"manila-operator-controller-manager-65d89cfd9f-lqq2s\" (UID: \"d4376d0e-3428-49ec-993f-48b32d4c8863\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.565892 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-zrznv" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.569429 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.571754 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.572188 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4dzd\" (UniqueName: \"kubernetes.io/projected/5b816f82-7f2c-4537-aadd-53b6bded2e14-kube-api-access-c4dzd\") pod \"keystone-operator-controller-manager-655d88ccb9-t95db\" (UID: \"5b816f82-7f2c-4537-aadd-53b6bded2e14\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.572856 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.574637 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9xvp\" (UniqueName: \"kubernetes.io/projected/5603909a-7f2a-4c86-8446-9fae64c02482-kube-api-access-z9xvp\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t\" (UID: \"5603909a-7f2a-4c86-8446-9fae64c02482\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.574840 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.576651 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.588388 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-p99lt" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.592370 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.593413 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.594779 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.595197 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-mh9bv" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.596588 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.609081 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7bz7\" (UniqueName: \"kubernetes.io/projected/34735d37-aafa-4641-a951-f4d4d7fdab3b-kube-api-access-r7bz7\") pod \"nova-operator-controller-manager-7c7fc454ff-bg7zq\" (UID: \"34735d37-aafa-4641-a951-f4d4d7fdab3b\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.609128 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57sn9\" (UniqueName: \"kubernetes.io/projected/17114ec7-df18-4456-9e51-e93d74881841-kube-api-access-57sn9\") pod \"ovn-operator-controller-manager-579449c7d5-rq5fl\" (UID: \"17114ec7-df18-4456-9e51-e93d74881841\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.609192 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbhr5\" (UniqueName: \"kubernetes.io/projected/b457fc86-0c8f-4c98-9cbe-deddce25e44e-kube-api-access-rbhr5\") pod \"octavia-operator-controller-manager-7468f855d8-tf2wf\" (UID: \"b457fc86-0c8f-4c98-9cbe-deddce25e44e\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.609216 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24b8e975-2aeb-46ff-81e7-1b006bc37688-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g\" (UID: \"24b8e975-2aeb-46ff-81e7-1b006bc37688\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.609263 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd6v4\" (UniqueName: \"kubernetes.io/projected/24b8e975-2aeb-46ff-81e7-1b006bc37688-kube-api-access-hd6v4\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g\" (UID: \"24b8e975-2aeb-46ff-81e7-1b006bc37688\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.630716 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.640454 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.657966 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.661404 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7bz7\" (UniqueName: \"kubernetes.io/projected/34735d37-aafa-4641-a951-f4d4d7fdab3b-kube-api-access-r7bz7\") pod \"nova-operator-controller-manager-7c7fc454ff-bg7zq\" (UID: \"34735d37-aafa-4641-a951-f4d4d7fdab3b\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.663935 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.665096 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.667336 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-4j946" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.710283 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2s8j\" (UniqueName: \"kubernetes.io/projected/8e918774-159b-41dd-b320-5aaded1b8f52-kube-api-access-w2s8j\") pod \"placement-operator-controller-manager-54689d9f88-qxvcq\" (UID: \"8e918774-159b-41dd-b320-5aaded1b8f52\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.710343 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57sn9\" (UniqueName: \"kubernetes.io/projected/17114ec7-df18-4456-9e51-e93d74881841-kube-api-access-57sn9\") pod \"ovn-operator-controller-manager-579449c7d5-rq5fl\" (UID: \"17114ec7-df18-4456-9e51-e93d74881841\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.710401 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbhr5\" (UniqueName: \"kubernetes.io/projected/b457fc86-0c8f-4c98-9cbe-deddce25e44e-kube-api-access-rbhr5\") pod \"octavia-operator-controller-manager-7468f855d8-tf2wf\" (UID: \"b457fc86-0c8f-4c98-9cbe-deddce25e44e\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.710435 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjz9f\" (UniqueName: \"kubernetes.io/projected/5c506bb2-6e70-425c-b4ed-be10eb472389-kube-api-access-qjz9f\") pod \"swift-operator-controller-manager-6859f9b676-jznzg\" (UID: \"5c506bb2-6e70-425c-b4ed-be10eb472389\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.710458 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24b8e975-2aeb-46ff-81e7-1b006bc37688-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g\" (UID: \"24b8e975-2aeb-46ff-81e7-1b006bc37688\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:30 crc kubenswrapper[4969]: E1004 08:31:30.710971 4969 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 08:31:30 crc kubenswrapper[4969]: E1004 08:31:30.711034 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24b8e975-2aeb-46ff-81e7-1b006bc37688-cert podName:24b8e975-2aeb-46ff-81e7-1b006bc37688 nodeName:}" failed. No retries permitted until 2025-10-04 08:31:31.21101658 +0000 UTC m=+918.965285394 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/24b8e975-2aeb-46ff-81e7-1b006bc37688-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" (UID: "24b8e975-2aeb-46ff-81e7-1b006bc37688") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.715372 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.710487 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnffd\" (UniqueName: \"kubernetes.io/projected/d882bca8-0cc7-4053-8684-82707b083709-kube-api-access-nnffd\") pod \"telemetry-operator-controller-manager-5d4d74dd89-kt6j5\" (UID: \"d882bca8-0cc7-4053-8684-82707b083709\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.717287 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd6v4\" (UniqueName: \"kubernetes.io/projected/24b8e975-2aeb-46ff-81e7-1b006bc37688-kube-api-access-hd6v4\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g\" (UID: \"24b8e975-2aeb-46ff-81e7-1b006bc37688\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.732362 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.733739 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.737585 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-kzzd6" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.744919 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.745777 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.753427 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbhr5\" (UniqueName: \"kubernetes.io/projected/b457fc86-0c8f-4c98-9cbe-deddce25e44e-kube-api-access-rbhr5\") pod \"octavia-operator-controller-manager-7468f855d8-tf2wf\" (UID: \"b457fc86-0c8f-4c98-9cbe-deddce25e44e\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.756094 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57sn9\" (UniqueName: \"kubernetes.io/projected/17114ec7-df18-4456-9e51-e93d74881841-kube-api-access-57sn9\") pod \"ovn-operator-controller-manager-579449c7d5-rq5fl\" (UID: \"17114ec7-df18-4456-9e51-e93d74881841\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.757193 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd6v4\" (UniqueName: \"kubernetes.io/projected/24b8e975-2aeb-46ff-81e7-1b006bc37688-kube-api-access-hd6v4\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g\" (UID: \"24b8e975-2aeb-46ff-81e7-1b006bc37688\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.761782 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.769029 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-676569d79-24njs"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.770903 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.771291 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-676569d79-24njs"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.773029 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-q7g5z" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.818047 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8qfl\" (UniqueName: \"kubernetes.io/projected/a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9-kube-api-access-w8qfl\") pod \"watcher-operator-controller-manager-676569d79-24njs\" (UID: \"a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9\") " pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.818098 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjz9f\" (UniqueName: \"kubernetes.io/projected/5c506bb2-6e70-425c-b4ed-be10eb472389-kube-api-access-qjz9f\") pod \"swift-operator-controller-manager-6859f9b676-jznzg\" (UID: \"5c506bb2-6e70-425c-b4ed-be10eb472389\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.818151 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnffd\" (UniqueName: \"kubernetes.io/projected/d882bca8-0cc7-4053-8684-82707b083709-kube-api-access-nnffd\") pod \"telemetry-operator-controller-manager-5d4d74dd89-kt6j5\" (UID: \"d882bca8-0cc7-4053-8684-82707b083709\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.818179 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rchbh\" (UniqueName: \"kubernetes.io/projected/8bf2e62a-8649-4b9e-8557-84441113bbdc-kube-api-access-rchbh\") pod \"test-operator-controller-manager-5cd5cb47d7-ktq5d\" (UID: \"8bf2e62a-8649-4b9e-8557-84441113bbdc\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.818204 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2s8j\" (UniqueName: \"kubernetes.io/projected/8e918774-159b-41dd-b320-5aaded1b8f52-kube-api-access-w2s8j\") pod \"placement-operator-controller-manager-54689d9f88-qxvcq\" (UID: \"8e918774-159b-41dd-b320-5aaded1b8f52\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.818647 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.829060 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.832769 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.833589 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.843267 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-64fd9" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.843621 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.851359 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnffd\" (UniqueName: \"kubernetes.io/projected/d882bca8-0cc7-4053-8684-82707b083709-kube-api-access-nnffd\") pod \"telemetry-operator-controller-manager-5d4d74dd89-kt6j5\" (UID: \"d882bca8-0cc7-4053-8684-82707b083709\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.855187 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2s8j\" (UniqueName: \"kubernetes.io/projected/8e918774-159b-41dd-b320-5aaded1b8f52-kube-api-access-w2s8j\") pod \"placement-operator-controller-manager-54689d9f88-qxvcq\" (UID: \"8e918774-159b-41dd-b320-5aaded1b8f52\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.856627 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.870047 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.874795 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjz9f\" (UniqueName: \"kubernetes.io/projected/5c506bb2-6e70-425c-b4ed-be10eb472389-kube-api-access-qjz9f\") pod \"swift-operator-controller-manager-6859f9b676-jznzg\" (UID: \"5c506bb2-6e70-425c-b4ed-be10eb472389\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.875833 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.881299 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.881867 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.894684 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-4spwr" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.900746 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz"] Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.919941 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5523db3e-3b39-4bda-8885-898df0dcd5e7-cert\") pod \"openstack-operator-controller-manager-68dfdd7f5c-nl8g9\" (UID: \"5523db3e-3b39-4bda-8885-898df0dcd5e7\") " pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.920021 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rchbh\" (UniqueName: \"kubernetes.io/projected/8bf2e62a-8649-4b9e-8557-84441113bbdc-kube-api-access-rchbh\") pod \"test-operator-controller-manager-5cd5cb47d7-ktq5d\" (UID: \"8bf2e62a-8649-4b9e-8557-84441113bbdc\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.920072 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxmbp\" (UniqueName: \"kubernetes.io/projected/0b27d59e-dad0-402c-b7db-7d0dd87bd68e-kube-api-access-wxmbp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz\" (UID: \"0b27d59e-dad0-402c-b7db-7d0dd87bd68e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.920145 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phssp\" (UniqueName: \"kubernetes.io/projected/5523db3e-3b39-4bda-8885-898df0dcd5e7-kube-api-access-phssp\") pod \"openstack-operator-controller-manager-68dfdd7f5c-nl8g9\" (UID: \"5523db3e-3b39-4bda-8885-898df0dcd5e7\") " pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.920189 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89f37d97-0acc-4940-bd3a-733f4e6ed592-cert\") pod \"infra-operator-controller-manager-658588b8c9-zg6hn\" (UID: \"89f37d97-0acc-4940-bd3a-733f4e6ed592\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.920241 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8qfl\" (UniqueName: \"kubernetes.io/projected/a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9-kube-api-access-w8qfl\") pod \"watcher-operator-controller-manager-676569d79-24njs\" (UID: \"a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9\") " pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.925372 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.933141 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89f37d97-0acc-4940-bd3a-733f4e6ed592-cert\") pod \"infra-operator-controller-manager-658588b8c9-zg6hn\" (UID: \"89f37d97-0acc-4940-bd3a-733f4e6ed592\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.941980 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rchbh\" (UniqueName: \"kubernetes.io/projected/8bf2e62a-8649-4b9e-8557-84441113bbdc-kube-api-access-rchbh\") pod \"test-operator-controller-manager-5cd5cb47d7-ktq5d\" (UID: \"8bf2e62a-8649-4b9e-8557-84441113bbdc\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.948198 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" Oct 04 08:31:30 crc kubenswrapper[4969]: I1004 08:31:30.959540 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.006349 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8qfl\" (UniqueName: \"kubernetes.io/projected/a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9-kube-api-access-w8qfl\") pod \"watcher-operator-controller-manager-676569d79-24njs\" (UID: \"a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9\") " pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.021248 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5523db3e-3b39-4bda-8885-898df0dcd5e7-cert\") pod \"openstack-operator-controller-manager-68dfdd7f5c-nl8g9\" (UID: \"5523db3e-3b39-4bda-8885-898df0dcd5e7\") " pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.021318 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxmbp\" (UniqueName: \"kubernetes.io/projected/0b27d59e-dad0-402c-b7db-7d0dd87bd68e-kube-api-access-wxmbp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz\" (UID: \"0b27d59e-dad0-402c-b7db-7d0dd87bd68e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.021378 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phssp\" (UniqueName: \"kubernetes.io/projected/5523db3e-3b39-4bda-8885-898df0dcd5e7-kube-api-access-phssp\") pod \"openstack-operator-controller-manager-68dfdd7f5c-nl8g9\" (UID: \"5523db3e-3b39-4bda-8885-898df0dcd5e7\") " pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:31 crc kubenswrapper[4969]: E1004 08:31:31.021699 4969 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 04 08:31:31 crc kubenswrapper[4969]: E1004 08:31:31.021773 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5523db3e-3b39-4bda-8885-898df0dcd5e7-cert podName:5523db3e-3b39-4bda-8885-898df0dcd5e7 nodeName:}" failed. No retries permitted until 2025-10-04 08:31:31.521753535 +0000 UTC m=+919.276022349 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5523db3e-3b39-4bda-8885-898df0dcd5e7-cert") pod "openstack-operator-controller-manager-68dfdd7f5c-nl8g9" (UID: "5523db3e-3b39-4bda-8885-898df0dcd5e7") : secret "webhook-server-cert" not found Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.042092 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxmbp\" (UniqueName: \"kubernetes.io/projected/0b27d59e-dad0-402c-b7db-7d0dd87bd68e-kube-api-access-wxmbp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz\" (UID: \"0b27d59e-dad0-402c-b7db-7d0dd87bd68e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.045025 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phssp\" (UniqueName: \"kubernetes.io/projected/5523db3e-3b39-4bda-8885-898df0dcd5e7-kube-api-access-phssp\") pod \"openstack-operator-controller-manager-68dfdd7f5c-nl8g9\" (UID: \"5523db3e-3b39-4bda-8885-898df0dcd5e7\") " pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.076749 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.107439 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.142911 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.155143 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv"] Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.163945 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.226375 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.226819 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24b8e975-2aeb-46ff-81e7-1b006bc37688-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g\" (UID: \"24b8e975-2aeb-46ff-81e7-1b006bc37688\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:31 crc kubenswrapper[4969]: E1004 08:31:31.227024 4969 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 08:31:31 crc kubenswrapper[4969]: E1004 08:31:31.227142 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24b8e975-2aeb-46ff-81e7-1b006bc37688-cert podName:24b8e975-2aeb-46ff-81e7-1b006bc37688 nodeName:}" failed. No retries permitted until 2025-10-04 08:31:32.227097217 +0000 UTC m=+919.981366031 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/24b8e975-2aeb-46ff-81e7-1b006bc37688-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" (UID: "24b8e975-2aeb-46ff-81e7-1b006bc37688") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.343510 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg"] Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.361202 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x"] Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.399960 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" event={"ID":"b7837ab8-145d-4f08-b407-5f0325119fb1","Type":"ContainerStarted","Data":"63e230834bf3742a6f1929b8e819768c790049200173e7875b2e4da7ba94508c"} Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.494591 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s"] Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.501735 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t"] Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.531137 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5523db3e-3b39-4bda-8885-898df0dcd5e7-cert\") pod \"openstack-operator-controller-manager-68dfdd7f5c-nl8g9\" (UID: \"5523db3e-3b39-4bda-8885-898df0dcd5e7\") " pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.539956 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5523db3e-3b39-4bda-8885-898df0dcd5e7-cert\") pod \"openstack-operator-controller-manager-68dfdd7f5c-nl8g9\" (UID: \"5523db3e-3b39-4bda-8885-898df0dcd5e7\") " pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.700210 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn"] Oct 04 08:31:31 crc kubenswrapper[4969]: W1004 08:31:31.708015 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc4ed288_8f9e_44a4_abda_81b7b4dfd8ea.slice/crio-0e027705c6c9397e9834ddf25fcd7ce68124ec187e509f9ea9d14a5d175274a5 WatchSource:0}: Error finding container 0e027705c6c9397e9834ddf25fcd7ce68124ec187e509f9ea9d14a5d175274a5: Status 404 returned error can't find the container with id 0e027705c6c9397e9834ddf25fcd7ce68124ec187e509f9ea9d14a5d175274a5 Oct 04 08:31:31 crc kubenswrapper[4969]: W1004 08:31:31.710704 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7dc7acf_3666_46fd_99cd_8447344cd10c.slice/crio-148d65ba7f65d6103efe43c093364004c5ee7edee7198e0a36594b57790e7d08 WatchSource:0}: Error finding container 148d65ba7f65d6103efe43c093364004c5ee7edee7198e0a36594b57790e7d08: Status 404 returned error can't find the container with id 148d65ba7f65d6103efe43c093364004c5ee7edee7198e0a36594b57790e7d08 Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.711244 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885"] Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.782751 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.909692 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq"] Oct 04 08:31:31 crc kubenswrapper[4969]: W1004 08:31:31.920039 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedfa72c9_187e_4d6c_8aab_66cb36073282.slice/crio-707103a38e3ebb2fae0c4d0b8c7a0a901f0028a5ab246eeffc8e1d71ea60aa58 WatchSource:0}: Error finding container 707103a38e3ebb2fae0c4d0b8c7a0a901f0028a5ab246eeffc8e1d71ea60aa58: Status 404 returned error can't find the container with id 707103a38e3ebb2fae0c4d0b8c7a0a901f0028a5ab246eeffc8e1d71ea60aa58 Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.951554 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db"] Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.955406 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr"] Oct 04 08:31:31 crc kubenswrapper[4969]: W1004 08:31:31.959638 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b816f82_7f2c_4537_aadd_53b6bded2e14.slice/crio-249e100caa63ede6203f15d71927e111eb67914000c8251153ec22e87175bd2c WatchSource:0}: Error finding container 249e100caa63ede6203f15d71927e111eb67914000c8251153ec22e87175bd2c: Status 404 returned error can't find the container with id 249e100caa63ede6203f15d71927e111eb67914000c8251153ec22e87175bd2c Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.960524 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8"] Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.970723 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq"] Oct 04 08:31:31 crc kubenswrapper[4969]: I1004 08:31:31.975372 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn"] Oct 04 08:31:31 crc kubenswrapper[4969]: W1004 08:31:31.981302 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18413b49_54b1_4ac9_8225_4aa748d0e4f3.slice/crio-5c984c36ffb83028e641deade36cd9c069d08868febbca12802713a0f991fcd9 WatchSource:0}: Error finding container 5c984c36ffb83028e641deade36cd9c069d08868febbca12802713a0f991fcd9: Status 404 returned error can't find the container with id 5c984c36ffb83028e641deade36cd9c069d08868febbca12802713a0f991fcd9 Oct 04 08:31:31 crc kubenswrapper[4969]: W1004 08:31:31.987271 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89f37d97_0acc_4940_bd3a_733f4e6ed592.slice/crio-58342004ec8fcaf35722956b5bee82d509a2ef8f39f659fe9970b82aea48f4e5 WatchSource:0}: Error finding container 58342004ec8fcaf35722956b5bee82d509a2ef8f39f659fe9970b82aea48f4e5: Status 404 returned error can't find the container with id 58342004ec8fcaf35722956b5bee82d509a2ef8f39f659fe9970b82aea48f4e5 Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.244287 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24b8e975-2aeb-46ff-81e7-1b006bc37688-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g\" (UID: \"24b8e975-2aeb-46ff-81e7-1b006bc37688\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.247893 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24b8e975-2aeb-46ff-81e7-1b006bc37688-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g\" (UID: \"24b8e975-2aeb-46ff-81e7-1b006bc37688\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.316818 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5"] Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.333113 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz"] Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.333162 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg"] Oct 04 08:31:32 crc kubenswrapper[4969]: W1004 08:31:32.348944 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bf2e62a_8649_4b9e_8557_84441113bbdc.slice/crio-6b6187d0ce518ae2347c119ca34cf4a51cb85a0c73c1ee715ce03808d68c4c1f WatchSource:0}: Error finding container 6b6187d0ce518ae2347c119ca34cf4a51cb85a0c73c1ee715ce03808d68c4c1f: Status 404 returned error can't find the container with id 6b6187d0ce518ae2347c119ca34cf4a51cb85a0c73c1ee715ce03808d68c4c1f Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.349833 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf"] Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.352056 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rchbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-ktq5d_openstack-operators(8bf2e62a-8649-4b9e-8557-84441113bbdc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.356608 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl"] Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.358380 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d"] Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.358653 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rbhr5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7468f855d8-tf2wf_openstack-operators(b457fc86-0c8f-4c98-9cbe-deddce25e44e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.363823 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq"] Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.375270 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-676569d79-24njs"] Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.383636 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9"] Oct 04 08:31:32 crc kubenswrapper[4969]: W1004 08:31:32.401651 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5523db3e_3b39_4bda_8885_898df0dcd5e7.slice/crio-4281c5a3e9c08bba1e5191faa771d17ae22d09cae5e6928513177c9677768924 WatchSource:0}: Error finding container 4281c5a3e9c08bba1e5191faa771d17ae22d09cae5e6928513177c9677768924: Status 404 returned error can't find the container with id 4281c5a3e9c08bba1e5191faa771d17ae22d09cae5e6928513177c9677768924 Oct 04 08:31:32 crc kubenswrapper[4969]: W1004 08:31:32.403127 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c506bb2_6e70_425c_b4ed_be10eb472389.slice/crio-a764b20b31dfbb31b1b43243bff0f33083d664bd451e1923bed0d679b3cfc8ce WatchSource:0}: Error finding container a764b20b31dfbb31b1b43243bff0f33083d664bd451e1923bed0d679b3cfc8ce: Status 404 returned error can't find the container with id a764b20b31dfbb31b1b43243bff0f33083d664bd451e1923bed0d679b3cfc8ce Oct 04 08:31:32 crc kubenswrapper[4969]: W1004 08:31:32.404128 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda65cbcc8_1908_47d5_b4bc_561f6dd8c9f9.slice/crio-6a254b56bf12adc8a06207c3d2b18d9d21b2786c16f9d36fe7bf7e5b44ce7af0 WatchSource:0}: Error finding container 6a254b56bf12adc8a06207c3d2b18d9d21b2786c16f9d36fe7bf7e5b44ce7af0: Status 404 returned error can't find the container with id 6a254b56bf12adc8a06207c3d2b18d9d21b2786c16f9d36fe7bf7e5b44ce7af0 Oct 04 08:31:32 crc kubenswrapper[4969]: W1004 08:31:32.405891 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e918774_159b_41dd_b320_5aaded1b8f52.slice/crio-d22c49488dcdf284bb0bdb86155227ef4dc05f0a97414fb822f85090f45a1484 WatchSource:0}: Error finding container d22c49488dcdf284bb0bdb86155227ef4dc05f0a97414fb822f85090f45a1484: Status 404 returned error can't find the container with id d22c49488dcdf284bb0bdb86155227ef4dc05f0a97414fb822f85090f45a1484 Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.406915 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qjz9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6859f9b676-jznzg_openstack-operators(5c506bb2-6e70-425c-b4ed-be10eb472389): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.407318 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.83:5001/openstack-k8s-operators/watcher-operator:dbf7c7dbd2d1678a881e5b109f14353af0f418fc,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w8qfl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-676569d79-24njs_openstack-operators(a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.408201 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w2s8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-54689d9f88-qxvcq_openstack-operators(8e918774-159b-41dd-b320-5aaded1b8f52): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.409517 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" event={"ID":"0ce59178-f382-4436-be86-5bc4c5d4a9b0","Type":"ContainerStarted","Data":"b639e5e2167613beec5231ed766851805561b472e75cbd1093bc8e58135d23d1"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.410780 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" event={"ID":"89f37d97-0acc-4940-bd3a-733f4e6ed592","Type":"ContainerStarted","Data":"58342004ec8fcaf35722956b5bee82d509a2ef8f39f659fe9970b82aea48f4e5"} Oct 04 08:31:32 crc kubenswrapper[4969]: W1004 08:31:32.411399 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17114ec7_df18_4456_9e51_e93d74881841.slice/crio-edba534c505bcacb0093d7e14d70fca66101f57a936b9a4c9ab9fa92ce7612af WatchSource:0}: Error finding container edba534c505bcacb0093d7e14d70fca66101f57a936b9a4c9ab9fa92ce7612af: Status 404 returned error can't find the container with id edba534c505bcacb0093d7e14d70fca66101f57a936b9a4c9ab9fa92ce7612af Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.412062 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" event={"ID":"d7dc7acf-3666-46fd-99cd-8447344cd10c","Type":"ContainerStarted","Data":"148d65ba7f65d6103efe43c093364004c5ee7edee7198e0a36594b57790e7d08"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.414973 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.422150 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" event={"ID":"34735d37-aafa-4641-a951-f4d4d7fdab3b","Type":"ContainerStarted","Data":"8955b7df013980ce0d4d0b5ec8253e4d5247af7da546ef5c9c9a769272a3fd77"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.423483 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" event={"ID":"d4376d0e-3428-49ec-993f-48b32d4c8863","Type":"ContainerStarted","Data":"563c4ed5bfddf3a923b3c365c66cb9441f84eaf8674fc69d139e9ad1b1d22b40"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.426115 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" event={"ID":"5603909a-7f2a-4c86-8446-9fae64c02482","Type":"ContainerStarted","Data":"0fb9290f6bb3d1e9dc69a1c5edd6bd76c3f465b44ddeb302bf03fd79c0d5ced8"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.427432 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" event={"ID":"edfa72c9-187e-4d6c-8aab-66cb36073282","Type":"ContainerStarted","Data":"707103a38e3ebb2fae0c4d0b8c7a0a901f0028a5ab246eeffc8e1d71ea60aa58"} Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.429947 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-57sn9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-579449c7d5-rq5fl_openstack-operators(17114ec7-df18-4456-9e51-e93d74881841): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.430658 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" event={"ID":"d882bca8-0cc7-4053-8684-82707b083709","Type":"ContainerStarted","Data":"588f9798b6f99b8573150ba044ef6c84d55a10265a5377eee4283d4e52d5b9c4"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.433249 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" event={"ID":"0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f","Type":"ContainerStarted","Data":"7cb8ac4179bd767e73847d4622315e8cfe51356ac87a99791cec704bc61bb9f3"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.437528 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" event={"ID":"8bf2e62a-8649-4b9e-8557-84441113bbdc","Type":"ContainerStarted","Data":"6b6187d0ce518ae2347c119ca34cf4a51cb85a0c73c1ee715ce03808d68c4c1f"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.440079 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" event={"ID":"18413b49-54b1-4ac9-8225-4aa748d0e4f3","Type":"ContainerStarted","Data":"5c984c36ffb83028e641deade36cd9c069d08868febbca12802713a0f991fcd9"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.442556 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" event={"ID":"eb8a54aa-2ce0-400c-a06f-cc3513b1df30","Type":"ContainerStarted","Data":"b5e96ae376e9c885cfecbca7f3f651302f716c1ac5efd47292d53f9123e8468e"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.443470 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" event={"ID":"b457fc86-0c8f-4c98-9cbe-deddce25e44e","Type":"ContainerStarted","Data":"cc111f9075e162ab2090fc98223b7e07c4a530f392c3845e936371bc5e07ab07"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.446056 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" event={"ID":"5b816f82-7f2c-4537-aadd-53b6bded2e14","Type":"ContainerStarted","Data":"249e100caa63ede6203f15d71927e111eb67914000c8251153ec22e87175bd2c"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.449363 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz" event={"ID":"0b27d59e-dad0-402c-b7db-7d0dd87bd68e","Type":"ContainerStarted","Data":"e0e2f0182df4d1d942607c5528e4722f8e02f4b1a310b1775e6bc07a954562a9"} Oct 04 08:31:32 crc kubenswrapper[4969]: I1004 08:31:32.450850 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" event={"ID":"fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea","Type":"ContainerStarted","Data":"0e027705c6c9397e9834ddf25fcd7ce68124ec187e509f9ea9d14a5d175274a5"} Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.569797 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" podUID="5c506bb2-6e70-425c-b4ed-be10eb472389" Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.616597 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" podUID="8bf2e62a-8649-4b9e-8557-84441113bbdc" Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.824894 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" podUID="17114ec7-df18-4456-9e51-e93d74881841" Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.841653 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" podUID="a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9" Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.885717 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" podUID="b457fc86-0c8f-4c98-9cbe-deddce25e44e" Oct 04 08:31:32 crc kubenswrapper[4969]: E1004 08:31:32.885826 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" podUID="8e918774-159b-41dd-b320-5aaded1b8f52" Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.079620 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g"] Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.514676 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" event={"ID":"5c506bb2-6e70-425c-b4ed-be10eb472389","Type":"ContainerStarted","Data":"31fe8cd66efb7e4260af297e474a84782bba8cd2113f06e6f2aa38045aee420d"} Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.514717 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" event={"ID":"5c506bb2-6e70-425c-b4ed-be10eb472389","Type":"ContainerStarted","Data":"a764b20b31dfbb31b1b43243bff0f33083d664bd451e1923bed0d679b3cfc8ce"} Oct 04 08:31:33 crc kubenswrapper[4969]: E1004 08:31:33.516239 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" podUID="5c506bb2-6e70-425c-b4ed-be10eb472389" Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.521006 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" event={"ID":"b457fc86-0c8f-4c98-9cbe-deddce25e44e","Type":"ContainerStarted","Data":"6dfff77b21586b501dd1c5d46cbc000ba04681d2ee31ca6462deb28ae2079ac7"} Oct 04 08:31:33 crc kubenswrapper[4969]: E1004 08:31:33.522584 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" podUID="b457fc86-0c8f-4c98-9cbe-deddce25e44e" Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.528284 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" event={"ID":"17114ec7-df18-4456-9e51-e93d74881841","Type":"ContainerStarted","Data":"9ab4f2b0696fc94fa8371f43a223d347471b675b6d28288a8be1af8e4efab14e"} Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.528325 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" event={"ID":"17114ec7-df18-4456-9e51-e93d74881841","Type":"ContainerStarted","Data":"edba534c505bcacb0093d7e14d70fca66101f57a936b9a4c9ab9fa92ce7612af"} Oct 04 08:31:33 crc kubenswrapper[4969]: E1004 08:31:33.530040 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" podUID="17114ec7-df18-4456-9e51-e93d74881841" Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.531310 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" event={"ID":"a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9","Type":"ContainerStarted","Data":"f3a24aee604b7fa7c220dc1c8068facbaae627fa9dd838e8efc96386989eb5da"} Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.531347 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" event={"ID":"a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9","Type":"ContainerStarted","Data":"6a254b56bf12adc8a06207c3d2b18d9d21b2786c16f9d36fe7bf7e5b44ce7af0"} Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.537894 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" event={"ID":"5523db3e-3b39-4bda-8885-898df0dcd5e7","Type":"ContainerStarted","Data":"f307b9f78a1a112a163c931a31c9937853bc5bc08accaa8fa5876780cbb9a44e"} Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.537934 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" event={"ID":"5523db3e-3b39-4bda-8885-898df0dcd5e7","Type":"ContainerStarted","Data":"552705ee6715c08dc822edde54ff8975be78076f525ea82bc88089443e79000e"} Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.537951 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.537960 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" event={"ID":"5523db3e-3b39-4bda-8885-898df0dcd5e7","Type":"ContainerStarted","Data":"4281c5a3e9c08bba1e5191faa771d17ae22d09cae5e6928513177c9677768924"} Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.540741 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" event={"ID":"8bf2e62a-8649-4b9e-8557-84441113bbdc","Type":"ContainerStarted","Data":"0933e1f2a3dfaa3a9633bcc98ce7df7084e4457186949a3bb1ef1f5de2637934"} Oct 04 08:31:33 crc kubenswrapper[4969]: E1004 08:31:33.543799 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.83:5001/openstack-k8s-operators/watcher-operator:dbf7c7dbd2d1678a881e5b109f14353af0f418fc\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" podUID="a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9" Oct 04 08:31:33 crc kubenswrapper[4969]: E1004 08:31:33.544016 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" podUID="8bf2e62a-8649-4b9e-8557-84441113bbdc" Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.557160 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" event={"ID":"8e918774-159b-41dd-b320-5aaded1b8f52","Type":"ContainerStarted","Data":"db13e0110b62c6871e276ca14573345e1405e19cdb382fa9c4dd6308ee886e71"} Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.557203 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" event={"ID":"8e918774-159b-41dd-b320-5aaded1b8f52","Type":"ContainerStarted","Data":"d22c49488dcdf284bb0bdb86155227ef4dc05f0a97414fb822f85090f45a1484"} Oct 04 08:31:33 crc kubenswrapper[4969]: E1004 08:31:33.562143 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" podUID="8e918774-159b-41dd-b320-5aaded1b8f52" Oct 04 08:31:33 crc kubenswrapper[4969]: I1004 08:31:33.679672 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" podStartSLOduration=3.679653881 podStartE2EDuration="3.679653881s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:31:33.645786778 +0000 UTC m=+921.400055602" watchObservedRunningTime="2025-10-04 08:31:33.679653881 +0000 UTC m=+921.433922695" Oct 04 08:31:34 crc kubenswrapper[4969]: E1004 08:31:34.563959 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" podUID="17114ec7-df18-4456-9e51-e93d74881841" Oct 04 08:31:34 crc kubenswrapper[4969]: E1004 08:31:34.564198 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" podUID="8bf2e62a-8649-4b9e-8557-84441113bbdc" Oct 04 08:31:34 crc kubenswrapper[4969]: E1004 08:31:34.564533 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.83:5001/openstack-k8s-operators/watcher-operator:dbf7c7dbd2d1678a881e5b109f14353af0f418fc\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" podUID="a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9" Oct 04 08:31:34 crc kubenswrapper[4969]: E1004 08:31:34.564702 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" podUID="b457fc86-0c8f-4c98-9cbe-deddce25e44e" Oct 04 08:31:34 crc kubenswrapper[4969]: E1004 08:31:34.564761 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" podUID="5c506bb2-6e70-425c-b4ed-be10eb472389" Oct 04 08:31:34 crc kubenswrapper[4969]: E1004 08:31:34.565618 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" podUID="8e918774-159b-41dd-b320-5aaded1b8f52" Oct 04 08:31:34 crc kubenswrapper[4969]: W1004 08:31:34.703081 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24b8e975_2aeb_46ff_81e7_1b006bc37688.slice/crio-aa03968b0cb451d97f8532be88957bc27e83cf3cfb11bc6ffe9f9f76598235f0 WatchSource:0}: Error finding container aa03968b0cb451d97f8532be88957bc27e83cf3cfb11bc6ffe9f9f76598235f0: Status 404 returned error can't find the container with id aa03968b0cb451d97f8532be88957bc27e83cf3cfb11bc6ffe9f9f76598235f0 Oct 04 08:31:35 crc kubenswrapper[4969]: I1004 08:31:35.569828 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" event={"ID":"24b8e975-2aeb-46ff-81e7-1b006bc37688","Type":"ContainerStarted","Data":"aa03968b0cb451d97f8532be88957bc27e83cf3cfb11bc6ffe9f9f76598235f0"} Oct 04 08:31:41 crc kubenswrapper[4969]: I1004 08:31:41.790304 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-68dfdd7f5c-nl8g9" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.662926 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" event={"ID":"0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f","Type":"ContainerStarted","Data":"95b16bc8744c4ea529f062b3f079ef40fd0a42a8ce887cdff97c0d0d3a324143"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.663461 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.663472 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" event={"ID":"0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f","Type":"ContainerStarted","Data":"2654d76002b654aada1b8181193945a12cf7382a216cd0815fd99db90a3b4ba2"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.671813 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" event={"ID":"0ce59178-f382-4436-be86-5bc4c5d4a9b0","Type":"ContainerStarted","Data":"854fcf5e3c85a3606faaf51c9c5a85098d1e32d4f2f0c56655b913404ea5fcf3"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.677874 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" event={"ID":"5603909a-7f2a-4c86-8446-9fae64c02482","Type":"ContainerStarted","Data":"247e2c724221a3f226f90ca8861e515825fdc70e63d4cf5a1a577911fec4a1c4"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.686861 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" event={"ID":"89f37d97-0acc-4940-bd3a-733f4e6ed592","Type":"ContainerStarted","Data":"2db9471bb6c9e0468111c1eabad2a12c697a5d96d9333d3165115f0d087a3047"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.686908 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" event={"ID":"89f37d97-0acc-4940-bd3a-733f4e6ed592","Type":"ContainerStarted","Data":"18b18f767fb2d967d5cc32d2fef46b185c5f63aaada2e35e246beb50ed693b3d"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.687068 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.687961 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" podStartSLOduration=2.777246237 podStartE2EDuration="14.687939284s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.448708894 +0000 UTC m=+919.202977698" lastFinishedPulling="2025-10-04 08:31:43.359401931 +0000 UTC m=+931.113670745" observedRunningTime="2025-10-04 08:31:44.685639786 +0000 UTC m=+932.439908610" watchObservedRunningTime="2025-10-04 08:31:44.687939284 +0000 UTC m=+932.442208088" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.693703 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" event={"ID":"fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea","Type":"ContainerStarted","Data":"b908abf75e4212b85d57b19d433d439c400e893551994a67fdcc991fc93756df"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.703658 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" event={"ID":"18413b49-54b1-4ac9-8225-4aa748d0e4f3","Type":"ContainerStarted","Data":"717798f526e4ee9c907bec220fa664e729de62a1c6dd918b00e606df2c30f5d5"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.705257 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" podStartSLOduration=3.301083318 podStartE2EDuration="14.705243414s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.991202989 +0000 UTC m=+919.745471803" lastFinishedPulling="2025-10-04 08:31:43.395363085 +0000 UTC m=+931.149631899" observedRunningTime="2025-10-04 08:31:44.704891616 +0000 UTC m=+932.459160430" watchObservedRunningTime="2025-10-04 08:31:44.705243414 +0000 UTC m=+932.459512228" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.723655 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" event={"ID":"edfa72c9-187e-4d6c-8aab-66cb36073282","Type":"ContainerStarted","Data":"f5686afe0d5af0c7e4b446235eca772224f7e43534206f9722ff8104426c1f56"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.736533 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" event={"ID":"eb8a54aa-2ce0-400c-a06f-cc3513b1df30","Type":"ContainerStarted","Data":"65f0376e78f3aee2608bec3c55885ff5ad8269877be56939389978dd664da8a0"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.738971 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" event={"ID":"24b8e975-2aeb-46ff-81e7-1b006bc37688","Type":"ContainerStarted","Data":"429f2e0e0e445259706e34702d308175caf0f6709123e38831f55fbf8f623ad8"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.740622 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" event={"ID":"b7837ab8-145d-4f08-b407-5f0325119fb1","Type":"ContainerStarted","Data":"fc3cae8f95bfecb4d987aa84e66cd7583c55c1e20858c0fe50ef5217eeb79d87"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.740659 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" event={"ID":"b7837ab8-145d-4f08-b407-5f0325119fb1","Type":"ContainerStarted","Data":"2f27939a4c39d1e536b06db2981b0569eb7b619741c93a4cb861b35b0e89cf93"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.740762 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.741740 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz" event={"ID":"0b27d59e-dad0-402c-b7db-7d0dd87bd68e","Type":"ContainerStarted","Data":"d8c9d27b36eae96290d5a7b6f572279220d8e5613e7d723aae080629df104977"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.744017 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" event={"ID":"d4376d0e-3428-49ec-993f-48b32d4c8863","Type":"ContainerStarted","Data":"dd3a190a57214c1bc9b0965ac308b256ccae072fdadc469e0be0150b57e4e597"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.744042 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" event={"ID":"d4376d0e-3428-49ec-993f-48b32d4c8863","Type":"ContainerStarted","Data":"b7a75a8ce4160d571a20f6983e33695b20ee367a9a9a64b479f499e87c7be34b"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.744368 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.745477 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" event={"ID":"d7dc7acf-3666-46fd-99cd-8447344cd10c","Type":"ContainerStarted","Data":"1a03615f890b60a82dcb332778a88ef074f913c72cb66813a111c9ea0cdd5f67"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.751800 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" event={"ID":"d882bca8-0cc7-4053-8684-82707b083709","Type":"ContainerStarted","Data":"bcc9ff2df77d328a3d02e9947fd11acf642f33590140716a624d2969e3666a05"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.758879 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" event={"ID":"5b816f82-7f2c-4537-aadd-53b6bded2e14","Type":"ContainerStarted","Data":"6d473c5c07c0d5d06ba2a652b9945f913f1025c901a66a246a4c3ff79781d1ff"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.762978 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" event={"ID":"34735d37-aafa-4641-a951-f4d4d7fdab3b","Type":"ContainerStarted","Data":"6ab29ebefc2202a8b406ae51fb7ec8b4f816a5316bd61852466e16d16423d368"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.763015 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" event={"ID":"34735d37-aafa-4641-a951-f4d4d7fdab3b","Type":"ContainerStarted","Data":"a18c1b092c942fc55ee79d1d0b89d2035364329afac8686de13c92457a6bcb56"} Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.763405 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.776071 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" podStartSLOduration=2.787997354 podStartE2EDuration="14.776055196s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.336498611 +0000 UTC m=+919.090767425" lastFinishedPulling="2025-10-04 08:31:43.324556453 +0000 UTC m=+931.078825267" observedRunningTime="2025-10-04 08:31:44.766250452 +0000 UTC m=+932.520519286" watchObservedRunningTime="2025-10-04 08:31:44.776055196 +0000 UTC m=+932.530324000" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.827430 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" podStartSLOduration=3.489390685 podStartE2EDuration="14.827403225s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.981681602 +0000 UTC m=+919.735950416" lastFinishedPulling="2025-10-04 08:31:43.319694132 +0000 UTC m=+931.073962956" observedRunningTime="2025-10-04 08:31:44.799813388 +0000 UTC m=+932.554082212" watchObservedRunningTime="2025-10-04 08:31:44.827403225 +0000 UTC m=+932.581672039" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.840170 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" podStartSLOduration=3.103145639 podStartE2EDuration="14.840151552s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.58753934 +0000 UTC m=+919.341808154" lastFinishedPulling="2025-10-04 08:31:43.324545253 +0000 UTC m=+931.078814067" observedRunningTime="2025-10-04 08:31:44.820523953 +0000 UTC m=+932.574792767" watchObservedRunningTime="2025-10-04 08:31:44.840151552 +0000 UTC m=+932.594420366" Oct 04 08:31:44 crc kubenswrapper[4969]: I1004 08:31:44.845584 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz" podStartSLOduration=3.809111904 podStartE2EDuration="14.845565477s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:32.34394394 +0000 UTC m=+920.098212744" lastFinishedPulling="2025-10-04 08:31:43.380397503 +0000 UTC m=+931.134666317" observedRunningTime="2025-10-04 08:31:44.838721437 +0000 UTC m=+932.592990241" watchObservedRunningTime="2025-10-04 08:31:44.845565477 +0000 UTC m=+932.599834291" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.781109 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" event={"ID":"eb8a54aa-2ce0-400c-a06f-cc3513b1df30","Type":"ContainerStarted","Data":"9690652c041b87e998445b615f5acdac03ea46a3b21572747d2ec0e83a86eb96"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.781734 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.785933 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" event={"ID":"24b8e975-2aeb-46ff-81e7-1b006bc37688","Type":"ContainerStarted","Data":"3c02a7bd8249045b8f7e28981cd416cef3bac6005a647cb60bf1c94a2d231bdb"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.787175 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.790947 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" event={"ID":"5b816f82-7f2c-4537-aadd-53b6bded2e14","Type":"ContainerStarted","Data":"1f2b70a4d6afd90958ff001c6258da1af9bc44c0fc919d49a630567330abe7f3"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.791214 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.797787 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" event={"ID":"fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea","Type":"ContainerStarted","Data":"c16455d000474d7b228533ce43faee6064675ccd428ed8d0b4ddf3e34780b4fb"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.797971 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.813963 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" podStartSLOduration=3.9265105670000002 podStartE2EDuration="15.813935104s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.427183099 +0000 UTC m=+919.181451913" lastFinishedPulling="2025-10-04 08:31:43.314607636 +0000 UTC m=+931.068876450" observedRunningTime="2025-10-04 08:31:45.813049862 +0000 UTC m=+933.567318716" watchObservedRunningTime="2025-10-04 08:31:45.813935104 +0000 UTC m=+933.568203948" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.815463 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" event={"ID":"0ce59178-f382-4436-be86-5bc4c5d4a9b0","Type":"ContainerStarted","Data":"ee9848281ae88dbfe1825c7e978ca731529f0aca9f5a9b5267c393b5fb8858ae"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.815579 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.819611 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" event={"ID":"edfa72c9-187e-4d6c-8aab-66cb36073282","Type":"ContainerStarted","Data":"18d6953ab0b6e13d038f47f2d4ec45ca85768369fcbd32107066df40b8ce1a4e"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.819844 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.822568 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" event={"ID":"d7dc7acf-3666-46fd-99cd-8447344cd10c","Type":"ContainerStarted","Data":"4278540a64ab1c0067bbb4b5dac2a992ca15450d05ca088e0dfef732c3034e36"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.822859 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.825003 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" event={"ID":"18413b49-54b1-4ac9-8225-4aa748d0e4f3","Type":"ContainerStarted","Data":"d982e92f12bb3ab9a3ae54051aaa76c6c29f1e0dbff4aaf5fd36109cecf3bb34"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.825146 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.827944 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" event={"ID":"d882bca8-0cc7-4053-8684-82707b083709","Type":"ContainerStarted","Data":"1a25ebcf25268c277b30bd06bc879133de651a3e7afb35777647ac634220f576"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.828185 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.831541 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" event={"ID":"5603909a-7f2a-4c86-8446-9fae64c02482","Type":"ContainerStarted","Data":"0cd05bf72684b772537383b656e66e7a37b20916c6e0d634ef9a4eee71ca82b1"} Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.836771 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" podStartSLOduration=4.205256826 podStartE2EDuration="15.836747522s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.71207735 +0000 UTC m=+919.466346164" lastFinishedPulling="2025-10-04 08:31:43.343568036 +0000 UTC m=+931.097836860" observedRunningTime="2025-10-04 08:31:45.833296406 +0000 UTC m=+933.587565250" watchObservedRunningTime="2025-10-04 08:31:45.836747522 +0000 UTC m=+933.591016366" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.840765 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.863688 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" podStartSLOduration=4.431433047 podStartE2EDuration="15.863662292s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.962684779 +0000 UTC m=+919.716953593" lastFinishedPulling="2025-10-04 08:31:43.394913984 +0000 UTC m=+931.149182838" observedRunningTime="2025-10-04 08:31:45.861761585 +0000 UTC m=+933.616030429" watchObservedRunningTime="2025-10-04 08:31:45.863662292 +0000 UTC m=+933.617931146" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.906959 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" podStartSLOduration=7.216607992 podStartE2EDuration="15.906933209s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:34.704984807 +0000 UTC m=+922.459253621" lastFinishedPulling="2025-10-04 08:31:43.395310004 +0000 UTC m=+931.149578838" observedRunningTime="2025-10-04 08:31:45.896892079 +0000 UTC m=+933.651160903" watchObservedRunningTime="2025-10-04 08:31:45.906933209 +0000 UTC m=+933.661202063" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.921162 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" podStartSLOduration=4.149524848 podStartE2EDuration="15.921128342s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.563737217 +0000 UTC m=+919.318006031" lastFinishedPulling="2025-10-04 08:31:43.335340711 +0000 UTC m=+931.089609525" observedRunningTime="2025-10-04 08:31:45.913472272 +0000 UTC m=+933.667741126" watchObservedRunningTime="2025-10-04 08:31:45.921128342 +0000 UTC m=+933.675397176" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.931262 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" podStartSLOduration=4.542904073 podStartE2EDuration="15.931246755s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.974934774 +0000 UTC m=+919.729203588" lastFinishedPulling="2025-10-04 08:31:43.363277456 +0000 UTC m=+931.117546270" observedRunningTime="2025-10-04 08:31:45.927689556 +0000 UTC m=+933.681958380" watchObservedRunningTime="2025-10-04 08:31:45.931246755 +0000 UTC m=+933.685515589" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.958645 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" podStartSLOduration=4.905383245 podStartE2EDuration="15.958623476s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:32.342774121 +0000 UTC m=+920.097042935" lastFinishedPulling="2025-10-04 08:31:43.396014352 +0000 UTC m=+931.150283166" observedRunningTime="2025-10-04 08:31:45.94715082 +0000 UTC m=+933.701419654" watchObservedRunningTime="2025-10-04 08:31:45.958623476 +0000 UTC m=+933.712892310" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.976165 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" podStartSLOduration=4.450582203 podStartE2EDuration="15.976138582s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.922359605 +0000 UTC m=+919.676628419" lastFinishedPulling="2025-10-04 08:31:43.447915984 +0000 UTC m=+931.202184798" observedRunningTime="2025-10-04 08:31:45.96722301 +0000 UTC m=+933.721491834" watchObservedRunningTime="2025-10-04 08:31:45.976138582 +0000 UTC m=+933.730407416" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.981188 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" podStartSLOduration=4.344929194 podStartE2EDuration="15.981169367s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.717297781 +0000 UTC m=+919.471566595" lastFinishedPulling="2025-10-04 08:31:43.353537954 +0000 UTC m=+931.107806768" observedRunningTime="2025-10-04 08:31:45.979655839 +0000 UTC m=+933.733924653" watchObservedRunningTime="2025-10-04 08:31:45.981169367 +0000 UTC m=+933.735438191" Oct 04 08:31:45 crc kubenswrapper[4969]: I1004 08:31:45.999773 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" podStartSLOduration=4.579150822 podStartE2EDuration="15.999756329s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:31.983661381 +0000 UTC m=+919.737930195" lastFinishedPulling="2025-10-04 08:31:43.404266888 +0000 UTC m=+931.158535702" observedRunningTime="2025-10-04 08:31:45.997282748 +0000 UTC m=+933.751551572" watchObservedRunningTime="2025-10-04 08:31:45.999756329 +0000 UTC m=+933.754025163" Oct 04 08:31:47 crc kubenswrapper[4969]: I1004 08:31:47.854617 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" event={"ID":"8e918774-159b-41dd-b320-5aaded1b8f52","Type":"ContainerStarted","Data":"d63eeca510530ce22983ac85069e1ec6a2cc579ad58cfc6e0817300842cbe273"} Oct 04 08:31:47 crc kubenswrapper[4969]: I1004 08:31:47.855263 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" Oct 04 08:31:47 crc kubenswrapper[4969]: I1004 08:31:47.859470 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" event={"ID":"a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9","Type":"ContainerStarted","Data":"d48f0e5f0fcc146122b23ffa6c8d73964524afd2a1755055ecacc9c6eec08d5b"} Oct 04 08:31:47 crc kubenswrapper[4969]: I1004 08:31:47.859748 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" Oct 04 08:31:47 crc kubenswrapper[4969]: I1004 08:31:47.873289 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" podStartSLOduration=2.966003676 podStartE2EDuration="17.873255369s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:32.408079577 +0000 UTC m=+920.162348391" lastFinishedPulling="2025-10-04 08:31:47.31533126 +0000 UTC m=+935.069600084" observedRunningTime="2025-10-04 08:31:47.869058055 +0000 UTC m=+935.623326869" watchObservedRunningTime="2025-10-04 08:31:47.873255369 +0000 UTC m=+935.627524223" Oct 04 08:31:47 crc kubenswrapper[4969]: I1004 08:31:47.888664 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" podStartSLOduration=2.955270129 podStartE2EDuration="17.888654093s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:32.407180004 +0000 UTC m=+920.161448818" lastFinishedPulling="2025-10-04 08:31:47.340563968 +0000 UTC m=+935.094832782" observedRunningTime="2025-10-04 08:31:47.886093919 +0000 UTC m=+935.640362743" watchObservedRunningTime="2025-10-04 08:31:47.888654093 +0000 UTC m=+935.642922917" Oct 04 08:31:48 crc kubenswrapper[4969]: I1004 08:31:48.871270 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" event={"ID":"5c506bb2-6e70-425c-b4ed-be10eb472389","Type":"ContainerStarted","Data":"15c511b9e5e464a368e2c4d76b625983d1742b6ad4f5e603776cc4ffc6c81a78"} Oct 04 08:31:48 crc kubenswrapper[4969]: I1004 08:31:48.871693 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" Oct 04 08:31:48 crc kubenswrapper[4969]: I1004 08:31:48.890772 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" podStartSLOduration=3.037443505 podStartE2EDuration="18.890757779s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:32.406791935 +0000 UTC m=+920.161060749" lastFinishedPulling="2025-10-04 08:31:48.260106209 +0000 UTC m=+936.014375023" observedRunningTime="2025-10-04 08:31:48.888231935 +0000 UTC m=+936.642500779" watchObservedRunningTime="2025-10-04 08:31:48.890757779 +0000 UTC m=+936.645026593" Oct 04 08:31:49 crc kubenswrapper[4969]: I1004 08:31:49.666095 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:31:49 crc kubenswrapper[4969]: I1004 08:31:49.666442 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:31:49 crc kubenswrapper[4969]: I1004 08:31:49.886188 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" event={"ID":"8bf2e62a-8649-4b9e-8557-84441113bbdc","Type":"ContainerStarted","Data":"ae7d965df7ad6815b48fb2e2c5327be2b4cbbaded5e2cf1b2a0e6685ec95f404"} Oct 04 08:31:49 crc kubenswrapper[4969]: I1004 08:31:49.886550 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" Oct 04 08:31:49 crc kubenswrapper[4969]: I1004 08:31:49.902263 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" podStartSLOduration=3.145644658 podStartE2EDuration="19.902231129s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:32.351951799 +0000 UTC m=+920.106220613" lastFinishedPulling="2025-10-04 08:31:49.10853827 +0000 UTC m=+936.862807084" observedRunningTime="2025-10-04 08:31:49.900210828 +0000 UTC m=+937.654479642" watchObservedRunningTime="2025-10-04 08:31:49.902231129 +0000 UTC m=+937.656499983" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.388311 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-m6llv" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.388406 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-x7vqg" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.460049 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-lvh4x" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.574809 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-lqq2s" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.601093 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.634042 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-54zxr" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.666381 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-d7885" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.753142 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-htxsn" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.767374 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-m8tn8" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.822121 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-x8xxq" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.843875 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-bg7zq" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.858659 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t95db" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.910138 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" event={"ID":"17114ec7-df18-4456-9e51-e93d74881841","Type":"ContainerStarted","Data":"2e9a415afd5e61baa6f73535c10f7228b62f2882cdf4ecbe3625db078d33a152"} Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.910787 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" Oct 04 08:31:50 crc kubenswrapper[4969]: I1004 08:31:50.929691 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" podStartSLOduration=3.470296759 podStartE2EDuration="20.929670365s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:32.429823018 +0000 UTC m=+920.184091832" lastFinishedPulling="2025-10-04 08:31:49.889196614 +0000 UTC m=+937.643465438" observedRunningTime="2025-10-04 08:31:50.929321217 +0000 UTC m=+938.683590031" watchObservedRunningTime="2025-10-04 08:31:50.929670365 +0000 UTC m=+938.683939179" Oct 04 08:31:51 crc kubenswrapper[4969]: I1004 08:31:51.083959 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-zg6hn" Oct 04 08:31:51 crc kubenswrapper[4969]: I1004 08:31:51.115450 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kt6j5" Oct 04 08:31:52 crc kubenswrapper[4969]: I1004 08:31:52.429052 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g" Oct 04 08:31:52 crc kubenswrapper[4969]: I1004 08:31:52.924234 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" event={"ID":"b457fc86-0c8f-4c98-9cbe-deddce25e44e","Type":"ContainerStarted","Data":"e91e51892627782759fcaf13d579960d90d3acb47920212b6bd0d1c8d5711d52"} Oct 04 08:31:52 crc kubenswrapper[4969]: I1004 08:31:52.924489 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" Oct 04 08:31:52 crc kubenswrapper[4969]: I1004 08:31:52.944544 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" podStartSLOduration=3.383092158 podStartE2EDuration="22.944529433s" podCreationTimestamp="2025-10-04 08:31:30 +0000 UTC" firstStartedPulling="2025-10-04 08:31:32.358543504 +0000 UTC m=+920.112812318" lastFinishedPulling="2025-10-04 08:31:51.919980789 +0000 UTC m=+939.674249593" observedRunningTime="2025-10-04 08:31:52.940602037 +0000 UTC m=+940.694870861" watchObservedRunningTime="2025-10-04 08:31:52.944529433 +0000 UTC m=+940.698798247" Oct 04 08:32:00 crc kubenswrapper[4969]: I1004 08:32:00.886321 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-tf2wf" Oct 04 08:32:00 crc kubenswrapper[4969]: I1004 08:32:00.929023 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-rq5fl" Oct 04 08:32:00 crc kubenswrapper[4969]: I1004 08:32:00.970641 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-jznzg" Oct 04 08:32:00 crc kubenswrapper[4969]: I1004 08:32:00.970814 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-qxvcq" Oct 04 08:32:01 crc kubenswrapper[4969]: I1004 08:32:01.146220 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ktq5d" Oct 04 08:32:01 crc kubenswrapper[4969]: I1004 08:32:01.173809 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-676569d79-24njs" Oct 04 08:32:19 crc kubenswrapper[4969]: I1004 08:32:19.669176 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:32:19 crc kubenswrapper[4969]: I1004 08:32:19.669757 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.670337 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698c778d7-hxpzz"] Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.672285 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.680648 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.680670 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-l9ngf" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.680784 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.684058 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.692605 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698c778d7-hxpzz"] Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.747933 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6dcdf6f545-cbvkz"] Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.749453 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.751763 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.758253 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dcdf6f545-cbvkz"] Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.816583 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/790314fe-563f-4149-a058-74f7073c6895-config\") pod \"dnsmasq-dns-698c778d7-hxpzz\" (UID: \"790314fe-563f-4149-a058-74f7073c6895\") " pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.816686 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fqbh\" (UniqueName: \"kubernetes.io/projected/790314fe-563f-4149-a058-74f7073c6895-kube-api-access-4fqbh\") pod \"dnsmasq-dns-698c778d7-hxpzz\" (UID: \"790314fe-563f-4149-a058-74f7073c6895\") " pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.918494 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs5qc\" (UniqueName: \"kubernetes.io/projected/a3d61042-65af-4739-a3d6-b820d577de2f-kube-api-access-zs5qc\") pod \"dnsmasq-dns-6dcdf6f545-cbvkz\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.918558 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fqbh\" (UniqueName: \"kubernetes.io/projected/790314fe-563f-4149-a058-74f7073c6895-kube-api-access-4fqbh\") pod \"dnsmasq-dns-698c778d7-hxpzz\" (UID: \"790314fe-563f-4149-a058-74f7073c6895\") " pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.918614 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/790314fe-563f-4149-a058-74f7073c6895-config\") pod \"dnsmasq-dns-698c778d7-hxpzz\" (UID: \"790314fe-563f-4149-a058-74f7073c6895\") " pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.918634 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-config\") pod \"dnsmasq-dns-6dcdf6f545-cbvkz\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.918663 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-dns-svc\") pod \"dnsmasq-dns-6dcdf6f545-cbvkz\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.920334 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/790314fe-563f-4149-a058-74f7073c6895-config\") pod \"dnsmasq-dns-698c778d7-hxpzz\" (UID: \"790314fe-563f-4149-a058-74f7073c6895\") " pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.938675 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fqbh\" (UniqueName: \"kubernetes.io/projected/790314fe-563f-4149-a058-74f7073c6895-kube-api-access-4fqbh\") pod \"dnsmasq-dns-698c778d7-hxpzz\" (UID: \"790314fe-563f-4149-a058-74f7073c6895\") " pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:20 crc kubenswrapper[4969]: I1004 08:32:20.987527 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.024262 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-config\") pod \"dnsmasq-dns-6dcdf6f545-cbvkz\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.024323 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-dns-svc\") pod \"dnsmasq-dns-6dcdf6f545-cbvkz\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.024358 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs5qc\" (UniqueName: \"kubernetes.io/projected/a3d61042-65af-4739-a3d6-b820d577de2f-kube-api-access-zs5qc\") pod \"dnsmasq-dns-6dcdf6f545-cbvkz\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.025336 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-config\") pod \"dnsmasq-dns-6dcdf6f545-cbvkz\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.025841 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-dns-svc\") pod \"dnsmasq-dns-6dcdf6f545-cbvkz\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.059608 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs5qc\" (UniqueName: \"kubernetes.io/projected/a3d61042-65af-4739-a3d6-b820d577de2f-kube-api-access-zs5qc\") pod \"dnsmasq-dns-6dcdf6f545-cbvkz\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.068707 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.383846 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dcdf6f545-cbvkz"] Oct 04 08:32:21 crc kubenswrapper[4969]: W1004 08:32:21.394500 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3d61042_65af_4739_a3d6_b820d577de2f.slice/crio-55030c82761d2616975d68d072edf0488ff0bae41964bb6414defc62138890ed WatchSource:0}: Error finding container 55030c82761d2616975d68d072edf0488ff0bae41964bb6414defc62138890ed: Status 404 returned error can't find the container with id 55030c82761d2616975d68d072edf0488ff0bae41964bb6414defc62138890ed Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.397034 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 08:32:21 crc kubenswrapper[4969]: I1004 08:32:21.491764 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698c778d7-hxpzz"] Oct 04 08:32:21 crc kubenswrapper[4969]: W1004 08:32:21.527071 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod790314fe_563f_4149_a058_74f7073c6895.slice/crio-11bfce6842b12433fd76423967fcf0557a3b46802066a95775f93eb8652553bd WatchSource:0}: Error finding container 11bfce6842b12433fd76423967fcf0557a3b46802066a95775f93eb8652553bd: Status 404 returned error can't find the container with id 11bfce6842b12433fd76423967fcf0557a3b46802066a95775f93eb8652553bd Oct 04 08:32:22 crc kubenswrapper[4969]: I1004 08:32:22.210120 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" event={"ID":"a3d61042-65af-4739-a3d6-b820d577de2f","Type":"ContainerStarted","Data":"55030c82761d2616975d68d072edf0488ff0bae41964bb6414defc62138890ed"} Oct 04 08:32:22 crc kubenswrapper[4969]: I1004 08:32:22.212908 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698c778d7-hxpzz" event={"ID":"790314fe-563f-4149-a058-74f7073c6895","Type":"ContainerStarted","Data":"11bfce6842b12433fd76423967fcf0557a3b46802066a95775f93eb8652553bd"} Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.354636 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698c778d7-hxpzz"] Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.400136 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bcc77c96f-6jtkx"] Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.401284 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.410106 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bcc77c96f-6jtkx"] Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.578556 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-config\") pod \"dnsmasq-dns-6bcc77c96f-6jtkx\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.578651 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-dns-svc\") pod \"dnsmasq-dns-6bcc77c96f-6jtkx\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.578707 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wxfh\" (UniqueName: \"kubernetes.io/projected/886c9045-f58f-419e-bf8c-90f5ae42cba0-kube-api-access-7wxfh\") pod \"dnsmasq-dns-6bcc77c96f-6jtkx\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.629741 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dcdf6f545-cbvkz"] Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.646581 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75d676bbb7-lhbhj"] Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.647704 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.667951 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75d676bbb7-lhbhj"] Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.679723 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-config\") pod \"dnsmasq-dns-6bcc77c96f-6jtkx\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.679795 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-dns-svc\") pod \"dnsmasq-dns-6bcc77c96f-6jtkx\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.679843 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wxfh\" (UniqueName: \"kubernetes.io/projected/886c9045-f58f-419e-bf8c-90f5ae42cba0-kube-api-access-7wxfh\") pod \"dnsmasq-dns-6bcc77c96f-6jtkx\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.680937 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-config\") pod \"dnsmasq-dns-6bcc77c96f-6jtkx\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.681641 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-dns-svc\") pod \"dnsmasq-dns-6bcc77c96f-6jtkx\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.704171 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wxfh\" (UniqueName: \"kubernetes.io/projected/886c9045-f58f-419e-bf8c-90f5ae42cba0-kube-api-access-7wxfh\") pod \"dnsmasq-dns-6bcc77c96f-6jtkx\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.722541 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.780612 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dm84\" (UniqueName: \"kubernetes.io/projected/b853223c-adde-4fa9-add8-2fe57bb240dd-kube-api-access-5dm84\") pod \"dnsmasq-dns-75d676bbb7-lhbhj\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.781107 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-dns-svc\") pod \"dnsmasq-dns-75d676bbb7-lhbhj\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.781254 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-config\") pod \"dnsmasq-dns-75d676bbb7-lhbhj\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.883185 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dm84\" (UniqueName: \"kubernetes.io/projected/b853223c-adde-4fa9-add8-2fe57bb240dd-kube-api-access-5dm84\") pod \"dnsmasq-dns-75d676bbb7-lhbhj\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.883278 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-dns-svc\") pod \"dnsmasq-dns-75d676bbb7-lhbhj\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.883313 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-config\") pod \"dnsmasq-dns-75d676bbb7-lhbhj\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.884332 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-config\") pod \"dnsmasq-dns-75d676bbb7-lhbhj\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.884540 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-dns-svc\") pod \"dnsmasq-dns-75d676bbb7-lhbhj\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.904571 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dm84\" (UniqueName: \"kubernetes.io/projected/b853223c-adde-4fa9-add8-2fe57bb240dd-kube-api-access-5dm84\") pod \"dnsmasq-dns-75d676bbb7-lhbhj\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.946928 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bcc77c96f-6jtkx"] Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.961803 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.976234 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dff579849-k2lss"] Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.977393 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:24 crc kubenswrapper[4969]: I1004 08:32:24.987149 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dff579849-k2lss"] Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.086126 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-config\") pod \"dnsmasq-dns-5dff579849-k2lss\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.086175 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-dns-svc\") pod \"dnsmasq-dns-5dff579849-k2lss\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.086467 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59s5w\" (UniqueName: \"kubernetes.io/projected/da1d60cc-fcef-4418-a7d1-2ea245c0a274-kube-api-access-59s5w\") pod \"dnsmasq-dns-5dff579849-k2lss\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.190367 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59s5w\" (UniqueName: \"kubernetes.io/projected/da1d60cc-fcef-4418-a7d1-2ea245c0a274-kube-api-access-59s5w\") pod \"dnsmasq-dns-5dff579849-k2lss\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.190467 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-config\") pod \"dnsmasq-dns-5dff579849-k2lss\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.190489 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-dns-svc\") pod \"dnsmasq-dns-5dff579849-k2lss\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.191186 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-dns-svc\") pod \"dnsmasq-dns-5dff579849-k2lss\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.191894 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-config\") pod \"dnsmasq-dns-5dff579849-k2lss\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.193301 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bcc77c96f-6jtkx"] Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.212450 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59s5w\" (UniqueName: \"kubernetes.io/projected/da1d60cc-fcef-4418-a7d1-2ea245c0a274-kube-api-access-59s5w\") pod \"dnsmasq-dns-5dff579849-k2lss\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.301651 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.520116 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.521703 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.524003 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.524303 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.524469 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.524903 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.525001 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.525015 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.525128 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-s4n5z" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.538868 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.697963 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698027 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698072 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698114 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698143 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698291 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698344 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1547b86a-ce65-4135-a8a5-957017c2271c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698449 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-config-data\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698465 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698481 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7cmd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-kube-api-access-v7cmd\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.698533 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1547b86a-ce65-4135-a8a5-957017c2271c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.792045 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.793468 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.795534 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.795773 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-w9hfz" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.796624 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.796752 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.796837 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.796757 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.796752 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801238 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801293 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801318 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801348 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1547b86a-ce65-4135-a8a5-957017c2271c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801394 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801417 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7cmd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-kube-api-access-v7cmd\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801454 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-config-data\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801490 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1547b86a-ce65-4135-a8a5-957017c2271c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801541 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801601 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.801639 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.802635 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.802873 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.803042 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.803159 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.803199 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.803879 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-config-data\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.808158 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.813208 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1547b86a-ce65-4135-a8a5-957017c2271c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.817756 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.818070 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.818135 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1547b86a-ce65-4135-a8a5-957017c2271c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.836773 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7cmd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-kube-api-access-v7cmd\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.852227 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.859614 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.902922 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.902977 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eaebec4-b71b-409d-b91e-1993605ce209-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.903013 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.903042 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffm99\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-kube-api-access-ffm99\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.903110 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.903170 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eaebec4-b71b-409d-b91e-1993605ce209-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.903207 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.903236 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.903257 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.903288 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:25 crc kubenswrapper[4969]: I1004 08:32:25.903314 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004140 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004182 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004208 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004229 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004248 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004262 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eaebec4-b71b-409d-b91e-1993605ce209-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004285 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004303 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffm99\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-kube-api-access-ffm99\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004331 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004375 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eaebec4-b71b-409d-b91e-1993605ce209-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004399 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.004606 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.005250 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.005357 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.005580 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.005765 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.007197 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.009388 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eaebec4-b71b-409d-b91e-1993605ce209-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.010293 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.017350 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.020312 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffm99\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-kube-api-access-ffm99\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.021004 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eaebec4-b71b-409d-b91e-1993605ce209-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.033556 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.080328 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.081575 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.086827 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-notifications-svc" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.087058 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-default-user" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.087678 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-erlang-cookie" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.087791 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-plugins-conf" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.087859 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-server-conf" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.087897 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-server-dockercfg-kxdht" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.088053 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-config-data" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.100965 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.200547 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207213 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/28264a08-8168-419d-a20d-8f94a5c4ed77-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207254 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207280 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28264a08-8168-419d-a20d-8f94a5c4ed77-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207298 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207359 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207393 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6d92\" (UniqueName: \"kubernetes.io/projected/28264a08-8168-419d-a20d-8f94a5c4ed77-kube-api-access-g6d92\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207466 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207540 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/28264a08-8168-419d-a20d-8f94a5c4ed77-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207567 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/28264a08-8168-419d-a20d-8f94a5c4ed77-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207809 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/28264a08-8168-419d-a20d-8f94a5c4ed77-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.207868 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309214 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/28264a08-8168-419d-a20d-8f94a5c4ed77-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309278 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309339 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28264a08-8168-419d-a20d-8f94a5c4ed77-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309410 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309674 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309727 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6d92\" (UniqueName: \"kubernetes.io/projected/28264a08-8168-419d-a20d-8f94a5c4ed77-kube-api-access-g6d92\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309854 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309942 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/28264a08-8168-419d-a20d-8f94a5c4ed77-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309966 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/28264a08-8168-419d-a20d-8f94a5c4ed77-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.310010 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/28264a08-8168-419d-a20d-8f94a5c4ed77-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.310028 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.310084 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.310157 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.309941 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.310288 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28264a08-8168-419d-a20d-8f94a5c4ed77-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.310837 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/28264a08-8168-419d-a20d-8f94a5c4ed77-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.311630 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/28264a08-8168-419d-a20d-8f94a5c4ed77-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.313293 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.313673 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/28264a08-8168-419d-a20d-8f94a5c4ed77-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.319086 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/28264a08-8168-419d-a20d-8f94a5c4ed77-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.327574 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/28264a08-8168-419d-a20d-8f94a5c4ed77-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.329120 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.331819 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6d92\" (UniqueName: \"kubernetes.io/projected/28264a08-8168-419d-a20d-8f94a5c4ed77-kube-api-access-g6d92\") pod \"rabbitmq-notifications-server-0\" (UID: \"28264a08-8168-419d-a20d-8f94a5c4ed77\") " pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:26 crc kubenswrapper[4969]: I1004 08:32:26.407263 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.092211 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.093629 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.096451 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.096806 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-pzw4b" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.096867 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.096919 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.096999 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.101136 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.110569 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.239241 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.239310 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3d9463-8792-487d-b4a5-1b61ab3310f6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.239334 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a3d9463-8792-487d-b4a5-1b61ab3310f6-kolla-config\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.239354 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a3d9463-8792-487d-b4a5-1b61ab3310f6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.239370 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a3d9463-8792-487d-b4a5-1b61ab3310f6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.239401 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8a3d9463-8792-487d-b4a5-1b61ab3310f6-secrets\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.239495 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdv9v\" (UniqueName: \"kubernetes.io/projected/8a3d9463-8792-487d-b4a5-1b61ab3310f6-kube-api-access-hdv9v\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.239533 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a3d9463-8792-487d-b4a5-1b61ab3310f6-config-data-default\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.239573 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3d9463-8792-487d-b4a5-1b61ab3310f6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.340905 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdv9v\" (UniqueName: \"kubernetes.io/projected/8a3d9463-8792-487d-b4a5-1b61ab3310f6-kube-api-access-hdv9v\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.340958 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a3d9463-8792-487d-b4a5-1b61ab3310f6-config-data-default\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.340979 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3d9463-8792-487d-b4a5-1b61ab3310f6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.341008 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.341041 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3d9463-8792-487d-b4a5-1b61ab3310f6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.341060 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a3d9463-8792-487d-b4a5-1b61ab3310f6-kolla-config\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.341078 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a3d9463-8792-487d-b4a5-1b61ab3310f6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.341093 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a3d9463-8792-487d-b4a5-1b61ab3310f6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.341123 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8a3d9463-8792-487d-b4a5-1b61ab3310f6-secrets\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.341829 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a3d9463-8792-487d-b4a5-1b61ab3310f6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.341972 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.342042 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a3d9463-8792-487d-b4a5-1b61ab3310f6-config-data-default\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.342894 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a3d9463-8792-487d-b4a5-1b61ab3310f6-kolla-config\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.343544 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a3d9463-8792-487d-b4a5-1b61ab3310f6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.345944 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8a3d9463-8792-487d-b4a5-1b61ab3310f6-secrets\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.346214 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3d9463-8792-487d-b4a5-1b61ab3310f6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.358148 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdv9v\" (UniqueName: \"kubernetes.io/projected/8a3d9463-8792-487d-b4a5-1b61ab3310f6-kube-api-access-hdv9v\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.362858 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3d9463-8792-487d-b4a5-1b61ab3310f6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.376677 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"8a3d9463-8792-487d-b4a5-1b61ab3310f6\") " pod="openstack/openstack-galera-0" Oct 04 08:32:28 crc kubenswrapper[4969]: I1004 08:32:28.434717 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.014661 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.016149 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.021783 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.027132 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-47q5n" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.027353 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.028046 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.028280 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.168339 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfh76\" (UniqueName: \"kubernetes.io/projected/82fe815c-3f06-499f-9e3c-5a3b2f00d932-kube-api-access-xfh76\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.168488 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/82fe815c-3f06-499f-9e3c-5a3b2f00d932-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.168529 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/82fe815c-3f06-499f-9e3c-5a3b2f00d932-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.168572 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/82fe815c-3f06-499f-9e3c-5a3b2f00d932-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.168612 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.168638 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/82fe815c-3f06-499f-9e3c-5a3b2f00d932-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.168666 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/82fe815c-3f06-499f-9e3c-5a3b2f00d932-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.168988 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82fe815c-3f06-499f-9e3c-5a3b2f00d932-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.169032 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fe815c-3f06-499f-9e3c-5a3b2f00d932-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.270571 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.270684 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/82fe815c-3f06-499f-9e3c-5a3b2f00d932-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.270774 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/82fe815c-3f06-499f-9e3c-5a3b2f00d932-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.270889 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.270948 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82fe815c-3f06-499f-9e3c-5a3b2f00d932-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.271035 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fe815c-3f06-499f-9e3c-5a3b2f00d932-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.271117 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfh76\" (UniqueName: \"kubernetes.io/projected/82fe815c-3f06-499f-9e3c-5a3b2f00d932-kube-api-access-xfh76\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.271205 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/82fe815c-3f06-499f-9e3c-5a3b2f00d932-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.271266 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/82fe815c-3f06-499f-9e3c-5a3b2f00d932-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.271352 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/82fe815c-3f06-499f-9e3c-5a3b2f00d932-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.272240 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/82fe815c-3f06-499f-9e3c-5a3b2f00d932-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.273131 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/82fe815c-3f06-499f-9e3c-5a3b2f00d932-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.273253 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/82fe815c-3f06-499f-9e3c-5a3b2f00d932-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.277353 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82fe815c-3f06-499f-9e3c-5a3b2f00d932-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.280992 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/82fe815c-3f06-499f-9e3c-5a3b2f00d932-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.281222 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fe815c-3f06-499f-9e3c-5a3b2f00d932-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.281611 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/82fe815c-3f06-499f-9e3c-5a3b2f00d932-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.294884 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfh76\" (UniqueName: \"kubernetes.io/projected/82fe815c-3f06-499f-9e3c-5a3b2f00d932-kube-api-access-xfh76\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.303438 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"82fe815c-3f06-499f-9e3c-5a3b2f00d932\") " pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.351623 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.618029 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.619287 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: W1004 08:32:29.622885 4969 reflector.go:561] object-"openstack"/"memcached-memcached-dockercfg-nmwts": failed to list *v1.Secret: secrets "memcached-memcached-dockercfg-nmwts" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 04 08:32:29 crc kubenswrapper[4969]: E1004 08:32:29.622926 4969 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"memcached-memcached-dockercfg-nmwts\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"memcached-memcached-dockercfg-nmwts\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 08:32:29 crc kubenswrapper[4969]: W1004 08:32:29.622973 4969 reflector.go:561] object-"openstack"/"cert-memcached-svc": failed to list *v1.Secret: secrets "cert-memcached-svc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 04 08:32:29 crc kubenswrapper[4969]: E1004 08:32:29.622984 4969 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cert-memcached-svc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-memcached-svc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 08:32:29 crc kubenswrapper[4969]: W1004 08:32:29.624552 4969 reflector.go:561] object-"openstack"/"memcached-config-data": failed to list *v1.ConfigMap: configmaps "memcached-config-data" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 04 08:32:29 crc kubenswrapper[4969]: E1004 08:32:29.624576 4969 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"memcached-config-data\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"memcached-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.643247 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.679161 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-config-data\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.679206 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d6e66e-6079-4278-a26c-04a299dd6d9d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.679227 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t95wt\" (UniqueName: \"kubernetes.io/projected/41d6e66e-6079-4278-a26c-04a299dd6d9d-kube-api-access-t95wt\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.679340 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d6e66e-6079-4278-a26c-04a299dd6d9d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.679364 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-kolla-config\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.781106 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-kolla-config\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.781189 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-config-data\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.781213 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d6e66e-6079-4278-a26c-04a299dd6d9d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.781231 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t95wt\" (UniqueName: \"kubernetes.io/projected/41d6e66e-6079-4278-a26c-04a299dd6d9d-kube-api-access-t95wt\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.781291 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d6e66e-6079-4278-a26c-04a299dd6d9d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.785515 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d6e66e-6079-4278-a26c-04a299dd6d9d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:29 crc kubenswrapper[4969]: I1004 08:32:29.798326 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t95wt\" (UniqueName: \"kubernetes.io/projected/41d6e66e-6079-4278-a26c-04a299dd6d9d-kube-api-access-t95wt\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:30 crc kubenswrapper[4969]: I1004 08:32:30.274940 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" event={"ID":"886c9045-f58f-419e-bf8c-90f5ae42cba0","Type":"ContainerStarted","Data":"ecd3e4770bb36775daee41579a4fe1dce9e09dfb2dcc9c901ed8bf05b7f79f4c"} Oct 04 08:32:30 crc kubenswrapper[4969]: I1004 08:32:30.705268 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-nmwts" Oct 04 08:32:30 crc kubenswrapper[4969]: E1004 08:32:30.781706 4969 secret.go:188] Couldn't get secret openstack/cert-memcached-svc: failed to sync secret cache: timed out waiting for the condition Oct 04 08:32:30 crc kubenswrapper[4969]: E1004 08:32:30.781781 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41d6e66e-6079-4278-a26c-04a299dd6d9d-memcached-tls-certs podName:41d6e66e-6079-4278-a26c-04a299dd6d9d nodeName:}" failed. No retries permitted until 2025-10-04 08:32:31.281762329 +0000 UTC m=+979.036031153 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memcached-tls-certs" (UniqueName: "kubernetes.io/secret/41d6e66e-6079-4278-a26c-04a299dd6d9d-memcached-tls-certs") pod "memcached-0" (UID: "41d6e66e-6079-4278-a26c-04a299dd6d9d") : failed to sync secret cache: timed out waiting for the condition Oct 04 08:32:30 crc kubenswrapper[4969]: E1004 08:32:30.781815 4969 configmap.go:193] Couldn't get configMap openstack/memcached-config-data: failed to sync configmap cache: timed out waiting for the condition Oct 04 08:32:30 crc kubenswrapper[4969]: E1004 08:32:30.781955 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-kolla-config podName:41d6e66e-6079-4278-a26c-04a299dd6d9d nodeName:}" failed. No retries permitted until 2025-10-04 08:32:31.281920102 +0000 UTC m=+979.036188986 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kolla-config" (UniqueName: "kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-kolla-config") pod "memcached-0" (UID: "41d6e66e-6079-4278-a26c-04a299dd6d9d") : failed to sync configmap cache: timed out waiting for the condition Oct 04 08:32:30 crc kubenswrapper[4969]: E1004 08:32:30.781835 4969 configmap.go:193] Couldn't get configMap openstack/memcached-config-data: failed to sync configmap cache: timed out waiting for the condition Oct 04 08:32:30 crc kubenswrapper[4969]: E1004 08:32:30.782062 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-config-data podName:41d6e66e-6079-4278-a26c-04a299dd6d9d nodeName:}" failed. No retries permitted until 2025-10-04 08:32:31.282013736 +0000 UTC m=+979.036282590 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-config-data") pod "memcached-0" (UID: "41d6e66e-6079-4278-a26c-04a299dd6d9d") : failed to sync configmap cache: timed out waiting for the condition Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.017691 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.158490 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.310773 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d6e66e-6079-4278-a26c-04a299dd6d9d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.310824 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-kolla-config\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.310875 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-config-data\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.311770 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-config-data\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.311775 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41d6e66e-6079-4278-a26c-04a299dd6d9d-kolla-config\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.314401 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d6e66e-6079-4278-a26c-04a299dd6d9d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"41d6e66e-6079-4278-a26c-04a299dd6d9d\") " pod="openstack/memcached-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.434034 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.659245 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.660190 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.667024 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-nbzzn" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.711674 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.819677 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbdbm\" (UniqueName: \"kubernetes.io/projected/0e92558f-4f16-443e-b9a7-1c7f76013d2f-kube-api-access-pbdbm\") pod \"kube-state-metrics-0\" (UID: \"0e92558f-4f16-443e-b9a7-1c7f76013d2f\") " pod="openstack/kube-state-metrics-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.920715 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbdbm\" (UniqueName: \"kubernetes.io/projected/0e92558f-4f16-443e-b9a7-1c7f76013d2f-kube-api-access-pbdbm\") pod \"kube-state-metrics-0\" (UID: \"0e92558f-4f16-443e-b9a7-1c7f76013d2f\") " pod="openstack/kube-state-metrics-0" Oct 04 08:32:31 crc kubenswrapper[4969]: I1004 08:32:31.937017 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbdbm\" (UniqueName: \"kubernetes.io/projected/0e92558f-4f16-443e-b9a7-1c7f76013d2f-kube-api-access-pbdbm\") pod \"kube-state-metrics-0\" (UID: \"0e92558f-4f16-443e-b9a7-1c7f76013d2f\") " pod="openstack/kube-state-metrics-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.007972 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.856031 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.859465 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.874538 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-n8xt6" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.874885 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.876673 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.877249 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.877567 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.877826 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.895271 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.936028 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.936120 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/49069f68-580b-41f3-9504-6e8b722459f7-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.936210 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.936264 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.936291 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-config\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.936362 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.936475 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbcjh\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-kube-api-access-vbcjh\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:32 crc kubenswrapper[4969]: I1004 08:32:32.936538 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/49069f68-580b-41f3-9504-6e8b722459f7-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.037722 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.039460 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/49069f68-580b-41f3-9504-6e8b722459f7-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.039640 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.039707 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.039744 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-config\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.039804 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.039936 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbcjh\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-kube-api-access-vbcjh\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.039983 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/49069f68-580b-41f3-9504-6e8b722459f7-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.041878 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/49069f68-580b-41f3-9504-6e8b722459f7-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.042905 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.044599 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.047890 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-config\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.053068 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/49069f68-580b-41f3-9504-6e8b722459f7-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.053188 4969 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.053214 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6bc02e5d9caae1a8b7f65724c18b281ae808553b7b037139e0479e345eca79b6/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.059773 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.069094 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbcjh\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-kube-api-access-vbcjh\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.092012 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.190465 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 08:32:33 crc kubenswrapper[4969]: I1004 08:32:33.388865 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.803120 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fms7j"] Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.805518 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.808935 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.808996 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-pd2xb" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.809108 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-p9gfr"] Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.810644 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.827205 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fms7j"] Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.828080 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.873364 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-p9gfr"] Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.876348 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-var-run-ovn\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.876414 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-var-log-ovn\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.876508 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6d0bb5-43f9-4c31-838b-f71927d68b08-scripts\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.876539 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-ovn-controller-tls-certs\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.876659 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-var-run\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.876676 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-combined-ca-bundle\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.876690 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzjj8\" (UniqueName: \"kubernetes.io/projected/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-kube-api-access-zzjj8\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.876738 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-scripts\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.876770 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6lpz\" (UniqueName: \"kubernetes.io/projected/3f6d0bb5-43f9-4c31-838b-f71927d68b08-kube-api-access-j6lpz\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.877041 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-var-run\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.877177 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-var-lib\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.877218 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-etc-ovs\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.877276 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-var-log\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.978813 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6lpz\" (UniqueName: \"kubernetes.io/projected/3f6d0bb5-43f9-4c31-838b-f71927d68b08-kube-api-access-j6lpz\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.978896 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-var-run\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.978926 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-var-lib\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.978947 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-etc-ovs\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.978963 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-var-log\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.979331 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-var-run-ovn\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.979620 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-var-log-ovn\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.979647 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6d0bb5-43f9-4c31-838b-f71927d68b08-scripts\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.979646 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-var-log\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.979581 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-etc-ovs\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.979559 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-var-lib\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.979737 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-var-log-ovn\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.979764 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f6d0bb5-43f9-4c31-838b-f71927d68b08-var-run\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.980641 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-var-run-ovn\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.980956 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-ovn-controller-tls-certs\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.981129 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-var-run\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.981199 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-combined-ca-bundle\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.981223 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzjj8\" (UniqueName: \"kubernetes.io/projected/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-kube-api-access-zzjj8\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.981247 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-scripts\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.984094 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6d0bb5-43f9-4c31-838b-f71927d68b08-scripts\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.985223 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-scripts\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.985582 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-var-run\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.986820 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-ovn-controller-tls-certs\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:34 crc kubenswrapper[4969]: I1004 08:32:34.988545 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-combined-ca-bundle\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.005677 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6lpz\" (UniqueName: \"kubernetes.io/projected/3f6d0bb5-43f9-4c31-838b-f71927d68b08-kube-api-access-j6lpz\") pod \"ovn-controller-ovs-p9gfr\" (UID: \"3f6d0bb5-43f9-4c31-838b-f71927d68b08\") " pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.009775 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzjj8\" (UniqueName: \"kubernetes.io/projected/452b4c0a-fcb1-48ed-8c3a-87c34904a0a5-kube-api-access-zzjj8\") pod \"ovn-controller-fms7j\" (UID: \"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5\") " pod="openstack/ovn-controller-fms7j" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.144379 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fms7j" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.161728 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.341393 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.342916 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.344744 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.345118 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.345319 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.345512 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-cbskp" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.345806 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.348201 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.386401 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2f95cc-1c69-4731-90a9-a57b693b87ac-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.386464 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.386502 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-968rp\" (UniqueName: \"kubernetes.io/projected/8d2f95cc-1c69-4731-90a9-a57b693b87ac-kube-api-access-968rp\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.386531 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2f95cc-1c69-4731-90a9-a57b693b87ac-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.386555 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2f95cc-1c69-4731-90a9-a57b693b87ac-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.386742 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8d2f95cc-1c69-4731-90a9-a57b693b87ac-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.386903 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d2f95cc-1c69-4731-90a9-a57b693b87ac-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.386969 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2f95cc-1c69-4731-90a9-a57b693b87ac-config\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.488029 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2f95cc-1c69-4731-90a9-a57b693b87ac-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.488097 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.488141 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-968rp\" (UniqueName: \"kubernetes.io/projected/8d2f95cc-1c69-4731-90a9-a57b693b87ac-kube-api-access-968rp\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.488177 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2f95cc-1c69-4731-90a9-a57b693b87ac-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.488207 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2f95cc-1c69-4731-90a9-a57b693b87ac-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.488259 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8d2f95cc-1c69-4731-90a9-a57b693b87ac-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.488294 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d2f95cc-1c69-4731-90a9-a57b693b87ac-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.488317 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2f95cc-1c69-4731-90a9-a57b693b87ac-config\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.489949 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2f95cc-1c69-4731-90a9-a57b693b87ac-config\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.490728 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.492967 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8d2f95cc-1c69-4731-90a9-a57b693b87ac-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.493267 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d2f95cc-1c69-4731-90a9-a57b693b87ac-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.494800 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2f95cc-1c69-4731-90a9-a57b693b87ac-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.497136 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2f95cc-1c69-4731-90a9-a57b693b87ac-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.497537 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2f95cc-1c69-4731-90a9-a57b693b87ac-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.514791 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.515583 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-968rp\" (UniqueName: \"kubernetes.io/projected/8d2f95cc-1c69-4731-90a9-a57b693b87ac-kube-api-access-968rp\") pod \"ovsdbserver-nb-0\" (UID: \"8d2f95cc-1c69-4731-90a9-a57b693b87ac\") " pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:35 crc kubenswrapper[4969]: I1004 08:32:35.670485 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.387474 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.391087 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.393787 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8tc5s" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.394026 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.394142 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.394263 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.400127 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.548064 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.548108 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.548177 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-config\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.548196 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.548299 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.548347 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5fh6\" (UniqueName: \"kubernetes.io/projected/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-kube-api-access-k5fh6\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.548503 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.548572 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.649899 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-config\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.649943 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.650000 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.650858 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-config\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.650862 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.651049 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5fh6\" (UniqueName: \"kubernetes.io/projected/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-kube-api-access-k5fh6\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.651092 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.651123 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.651188 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.651212 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.651373 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.652089 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.663348 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.663842 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.665472 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.669288 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5fh6\" (UniqueName: \"kubernetes.io/projected/d179b5df-13f6-4e3f-bbd5-d210ba6b2f88-kube-api-access-k5fh6\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.686515 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88\") " pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:38 crc kubenswrapper[4969]: I1004 08:32:38.712792 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 08:32:39 crc kubenswrapper[4969]: I1004 08:32:39.255843 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75d676bbb7-lhbhj"] Oct 04 08:32:39 crc kubenswrapper[4969]: I1004 08:32:39.324903 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dff579849-k2lss"] Oct 04 08:32:39 crc kubenswrapper[4969]: I1004 08:32:39.351215 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8a3d9463-8792-487d-b4a5-1b61ab3310f6","Type":"ContainerStarted","Data":"cee3ee947df5f857492317d8b87416dc1fbb256149200325ed4f24bc772413a3"} Oct 04 08:32:40 crc kubenswrapper[4969]: W1004 08:32:40.016090 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb853223c_adde_4fa9_add8_2fe57bb240dd.slice/crio-101be2359ef0f9e71951ea383ae5075252004773281afa7ff4f0519691d704da WatchSource:0}: Error finding container 101be2359ef0f9e71951ea383ae5075252004773281afa7ff4f0519691d704da: Status 404 returned error can't find the container with id 101be2359ef0f9e71951ea383ae5075252004773281afa7ff4f0519691d704da Oct 04 08:32:40 crc kubenswrapper[4969]: E1004 08:32:40.046410 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 04 08:32:40 crc kubenswrapper[4969]: E1004 08:32:40.046912 4969 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 04 08:32:40 crc kubenswrapper[4969]: E1004 08:32:40.047333 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.83:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zs5qc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6dcdf6f545-cbvkz_openstack(a3d61042-65af-4739-a3d6-b820d577de2f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 08:32:40 crc kubenswrapper[4969]: E1004 08:32:40.048830 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" podUID="a3d61042-65af-4739-a3d6-b820d577de2f" Oct 04 08:32:40 crc kubenswrapper[4969]: W1004 08:32:40.051931 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda1d60cc_fcef_4418_a7d1_2ea245c0a274.slice/crio-c9bca4ec707e534dff9efe812ce3d823b7d9db793d4e5c49efa9b5f3b2cae69d WatchSource:0}: Error finding container c9bca4ec707e534dff9efe812ce3d823b7d9db793d4e5c49efa9b5f3b2cae69d: Status 404 returned error can't find the container with id c9bca4ec707e534dff9efe812ce3d823b7d9db793d4e5c49efa9b5f3b2cae69d Oct 04 08:32:40 crc kubenswrapper[4969]: E1004 08:32:40.058975 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 04 08:32:40 crc kubenswrapper[4969]: E1004 08:32:40.059027 4969 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 04 08:32:40 crc kubenswrapper[4969]: E1004 08:32:40.059171 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.83:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4fqbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-698c778d7-hxpzz_openstack(790314fe-563f-4149-a058-74f7073c6895): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 08:32:40 crc kubenswrapper[4969]: E1004 08:32:40.060355 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-698c778d7-hxpzz" podUID="790314fe-563f-4149-a058-74f7073c6895" Oct 04 08:32:40 crc kubenswrapper[4969]: I1004 08:32:40.362019 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dff579849-k2lss" event={"ID":"da1d60cc-fcef-4418-a7d1-2ea245c0a274","Type":"ContainerStarted","Data":"c9bca4ec707e534dff9efe812ce3d823b7d9db793d4e5c49efa9b5f3b2cae69d"} Oct 04 08:32:40 crc kubenswrapper[4969]: I1004 08:32:40.363298 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" event={"ID":"b853223c-adde-4fa9-add8-2fe57bb240dd","Type":"ContainerStarted","Data":"101be2359ef0f9e71951ea383ae5075252004773281afa7ff4f0519691d704da"} Oct 04 08:32:40 crc kubenswrapper[4969]: I1004 08:32:40.767844 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 08:32:40 crc kubenswrapper[4969]: I1004 08:32:40.780967 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 08:32:40 crc kubenswrapper[4969]: W1004 08:32:40.783529 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1547b86a_ce65_4135_a8a5_957017c2271c.slice/crio-72cf44fdbaf1b0c34f3c10eca9aa9599f6773c839e258d471b1a9c4532e0689c WatchSource:0}: Error finding container 72cf44fdbaf1b0c34f3c10eca9aa9599f6773c839e258d471b1a9c4532e0689c: Status 404 returned error can't find the container with id 72cf44fdbaf1b0c34f3c10eca9aa9599f6773c839e258d471b1a9c4532e0689c Oct 04 08:32:40 crc kubenswrapper[4969]: I1004 08:32:40.786860 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 08:32:40 crc kubenswrapper[4969]: W1004 08:32:40.794519 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6eaebec4_b71b_409d_b91e_1993605ce209.slice/crio-1e2b8014fdc03093198f115eb7098c67c5d21c3d5d9ba5d3b773ba5a903d006f WatchSource:0}: Error finding container 1e2b8014fdc03093198f115eb7098c67c5d21c3d5d9ba5d3b773ba5a903d006f: Status 404 returned error can't find the container with id 1e2b8014fdc03093198f115eb7098c67c5d21c3d5d9ba5d3b773ba5a903d006f Oct 04 08:32:40 crc kubenswrapper[4969]: W1004 08:32:40.800103 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82fe815c_3f06_499f_9e3c_5a3b2f00d932.slice/crio-1f199c85c15b61e00fff75764bf00d2e1757806f83016f5cbbd2707d0c6d8608 WatchSource:0}: Error finding container 1f199c85c15b61e00fff75764bf00d2e1757806f83016f5cbbd2707d0c6d8608: Status 404 returned error can't find the container with id 1f199c85c15b61e00fff75764bf00d2e1757806f83016f5cbbd2707d0c6d8608 Oct 04 08:32:40 crc kubenswrapper[4969]: I1004 08:32:40.884901 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:40 crc kubenswrapper[4969]: I1004 08:32:40.894968 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.001012 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.020740 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.034648 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-config\") pod \"a3d61042-65af-4739-a3d6-b820d577de2f\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.034715 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs5qc\" (UniqueName: \"kubernetes.io/projected/a3d61042-65af-4739-a3d6-b820d577de2f-kube-api-access-zs5qc\") pod \"a3d61042-65af-4739-a3d6-b820d577de2f\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.035731 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-config" (OuterVolumeSpecName: "config") pod "a3d61042-65af-4739-a3d6-b820d577de2f" (UID: "a3d61042-65af-4739-a3d6-b820d577de2f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.046394 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fqbh\" (UniqueName: \"kubernetes.io/projected/790314fe-563f-4149-a058-74f7073c6895-kube-api-access-4fqbh\") pod \"790314fe-563f-4149-a058-74f7073c6895\" (UID: \"790314fe-563f-4149-a058-74f7073c6895\") " Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.050051 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-dns-svc\") pod \"a3d61042-65af-4739-a3d6-b820d577de2f\" (UID: \"a3d61042-65af-4739-a3d6-b820d577de2f\") " Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.050396 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/790314fe-563f-4149-a058-74f7073c6895-config\") pod \"790314fe-563f-4149-a058-74f7073c6895\" (UID: \"790314fe-563f-4149-a058-74f7073c6895\") " Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.051789 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.049240 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3d61042-65af-4739-a3d6-b820d577de2f-kube-api-access-zs5qc" (OuterVolumeSpecName: "kube-api-access-zs5qc") pod "a3d61042-65af-4739-a3d6-b820d577de2f" (UID: "a3d61042-65af-4739-a3d6-b820d577de2f"). InnerVolumeSpecName "kube-api-access-zs5qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.052995 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790314fe-563f-4149-a058-74f7073c6895-config" (OuterVolumeSpecName: "config") pod "790314fe-563f-4149-a058-74f7073c6895" (UID: "790314fe-563f-4149-a058-74f7073c6895"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.053470 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a3d61042-65af-4739-a3d6-b820d577de2f" (UID: "a3d61042-65af-4739-a3d6-b820d577de2f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.046603 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fms7j"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.068979 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790314fe-563f-4149-a058-74f7073c6895-kube-api-access-4fqbh" (OuterVolumeSpecName: "kube-api-access-4fqbh") pod "790314fe-563f-4149-a058-74f7073c6895" (UID: "790314fe-563f-4149-a058-74f7073c6895"). InnerVolumeSpecName "kube-api-access-4fqbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:32:41 crc kubenswrapper[4969]: W1004 08:32:41.088665 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41d6e66e_6079_4278_a26c_04a299dd6d9d.slice/crio-8463fa18af1ad8a02a13539af3e53bbe3e6ecfd949a0dae18d5dab4bdd8fea1b WatchSource:0}: Error finding container 8463fa18af1ad8a02a13539af3e53bbe3e6ecfd949a0dae18d5dab4bdd8fea1b: Status 404 returned error can't find the container with id 8463fa18af1ad8a02a13539af3e53bbe3e6ecfd949a0dae18d5dab4bdd8fea1b Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.105522 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.105556 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.152052 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.152857 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d61042-65af-4739-a3d6-b820d577de2f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.152888 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/790314fe-563f-4149-a058-74f7073c6895-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.152898 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs5qc\" (UniqueName: \"kubernetes.io/projected/a3d61042-65af-4739-a3d6-b820d577de2f-kube-api-access-zs5qc\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.152909 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fqbh\" (UniqueName: \"kubernetes.io/projected/790314fe-563f-4149-a058-74f7073c6895-kube-api-access-4fqbh\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:41 crc kubenswrapper[4969]: W1004 08:32:41.157007 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd179b5df_13f6_4e3f_bbd5_d210ba6b2f88.slice/crio-6d68ff58ae500a0e3293618403336b1b97dd441a90a28662936e758f90c2ebbd WatchSource:0}: Error finding container 6d68ff58ae500a0e3293618403336b1b97dd441a90a28662936e758f90c2ebbd: Status 404 returned error can't find the container with id 6d68ff58ae500a0e3293618403336b1b97dd441a90a28662936e758f90c2ebbd Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.324295 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.374298 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerStarted","Data":"50089e4ac1d8024a9f1311d4600655ff4f6e53fade4e12109bc2d04d385edc65"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.376039 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88","Type":"ContainerStarted","Data":"6d68ff58ae500a0e3293618403336b1b97dd441a90a28662936e758f90c2ebbd"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.377676 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"28264a08-8168-419d-a20d-8f94a5c4ed77","Type":"ContainerStarted","Data":"aa70ebb1a06f0b5e03f09907d29017f16a916ac54b7daa861b80bc1658f5ff9a"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.378914 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1547b86a-ce65-4135-a8a5-957017c2271c","Type":"ContainerStarted","Data":"72cf44fdbaf1b0c34f3c10eca9aa9599f6773c839e258d471b1a9c4532e0689c"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.380044 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fms7j" event={"ID":"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5","Type":"ContainerStarted","Data":"3ec4a3d064d022d8ca99199d9306b6590d5e6b83244fce652f1449ce4f764bf2"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.382005 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.382705 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dcdf6f545-cbvkz" event={"ID":"a3d61042-65af-4739-a3d6-b820d577de2f","Type":"ContainerDied","Data":"55030c82761d2616975d68d072edf0488ff0bae41964bb6414defc62138890ed"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.389442 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698c778d7-hxpzz" event={"ID":"790314fe-563f-4149-a058-74f7073c6895","Type":"ContainerDied","Data":"11bfce6842b12433fd76423967fcf0557a3b46802066a95775f93eb8652553bd"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.389498 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698c778d7-hxpzz" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.391706 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0e92558f-4f16-443e-b9a7-1c7f76013d2f","Type":"ContainerStarted","Data":"97ca7e0f68581a2d2c39116e1db954233259dad3244f419260935f709390b9ff"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.394750 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"41d6e66e-6079-4278-a26c-04a299dd6d9d","Type":"ContainerStarted","Data":"8463fa18af1ad8a02a13539af3e53bbe3e6ecfd949a0dae18d5dab4bdd8fea1b"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.399402 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"82fe815c-3f06-499f-9e3c-5a3b2f00d932","Type":"ContainerStarted","Data":"1f199c85c15b61e00fff75764bf00d2e1757806f83016f5cbbd2707d0c6d8608"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.423842 4969 generic.go:334] "Generic (PLEG): container finished" podID="886c9045-f58f-419e-bf8c-90f5ae42cba0" containerID="ff0dbcc9381c93ef520fecd39929f2aecee9f297cec114c0253676d58e80ec02" exitCode=0 Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.424105 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" event={"ID":"886c9045-f58f-419e-bf8c-90f5ae42cba0","Type":"ContainerDied","Data":"ff0dbcc9381c93ef520fecd39929f2aecee9f297cec114c0253676d58e80ec02"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.437257 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dcdf6f545-cbvkz"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.449306 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6dcdf6f545-cbvkz"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.456476 4969 generic.go:334] "Generic (PLEG): container finished" podID="da1d60cc-fcef-4418-a7d1-2ea245c0a274" containerID="ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80" exitCode=0 Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.456567 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dff579849-k2lss" event={"ID":"da1d60cc-fcef-4418-a7d1-2ea245c0a274","Type":"ContainerDied","Data":"ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.461613 4969 generic.go:334] "Generic (PLEG): container finished" podID="b853223c-adde-4fa9-add8-2fe57bb240dd" containerID="7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4" exitCode=0 Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.461769 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" event={"ID":"b853223c-adde-4fa9-add8-2fe57bb240dd","Type":"ContainerDied","Data":"7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.471093 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698c778d7-hxpzz"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.476985 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698c778d7-hxpzz"] Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.485926 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6eaebec4-b71b-409d-b91e-1993605ce209","Type":"ContainerStarted","Data":"1e2b8014fdc03093198f115eb7098c67c5d21c3d5d9ba5d3b773ba5a903d006f"} Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.914686 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.964895 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wxfh\" (UniqueName: \"kubernetes.io/projected/886c9045-f58f-419e-bf8c-90f5ae42cba0-kube-api-access-7wxfh\") pod \"886c9045-f58f-419e-bf8c-90f5ae42cba0\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.965002 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-dns-svc\") pod \"886c9045-f58f-419e-bf8c-90f5ae42cba0\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.965151 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-config\") pod \"886c9045-f58f-419e-bf8c-90f5ae42cba0\" (UID: \"886c9045-f58f-419e-bf8c-90f5ae42cba0\") " Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.970908 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/886c9045-f58f-419e-bf8c-90f5ae42cba0-kube-api-access-7wxfh" (OuterVolumeSpecName: "kube-api-access-7wxfh") pod "886c9045-f58f-419e-bf8c-90f5ae42cba0" (UID: "886c9045-f58f-419e-bf8c-90f5ae42cba0"). InnerVolumeSpecName "kube-api-access-7wxfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.984129 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-config" (OuterVolumeSpecName: "config") pod "886c9045-f58f-419e-bf8c-90f5ae42cba0" (UID: "886c9045-f58f-419e-bf8c-90f5ae42cba0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:32:41 crc kubenswrapper[4969]: I1004 08:32:41.986839 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "886c9045-f58f-419e-bf8c-90f5ae42cba0" (UID: "886c9045-f58f-419e-bf8c-90f5ae42cba0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.060932 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-p9gfr"] Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.066797 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.066831 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wxfh\" (UniqueName: \"kubernetes.io/projected/886c9045-f58f-419e-bf8c-90f5ae42cba0-kube-api-access-7wxfh\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.066844 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886c9045-f58f-419e-bf8c-90f5ae42cba0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:42 crc kubenswrapper[4969]: W1004 08:32:42.158066 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f6d0bb5_43f9_4c31_838b_f71927d68b08.slice/crio-043521f19911a93dc718a36440dc90ac843a994d97bf383efd5c4cbe072945a2 WatchSource:0}: Error finding container 043521f19911a93dc718a36440dc90ac843a994d97bf383efd5c4cbe072945a2: Status 404 returned error can't find the container with id 043521f19911a93dc718a36440dc90ac843a994d97bf383efd5c4cbe072945a2 Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.494992 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.494967 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bcc77c96f-6jtkx" event={"ID":"886c9045-f58f-419e-bf8c-90f5ae42cba0","Type":"ContainerDied","Data":"ecd3e4770bb36775daee41579a4fe1dce9e09dfb2dcc9c901ed8bf05b7f79f4c"} Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.495065 4969 scope.go:117] "RemoveContainer" containerID="ff0dbcc9381c93ef520fecd39929f2aecee9f297cec114c0253676d58e80ec02" Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.497020 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p9gfr" event={"ID":"3f6d0bb5-43f9-4c31-838b-f71927d68b08","Type":"ContainerStarted","Data":"043521f19911a93dc718a36440dc90ac843a994d97bf383efd5c4cbe072945a2"} Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.498910 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dff579849-k2lss" event={"ID":"da1d60cc-fcef-4418-a7d1-2ea245c0a274","Type":"ContainerStarted","Data":"29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f"} Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.499053 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.503657 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" event={"ID":"b853223c-adde-4fa9-add8-2fe57bb240dd","Type":"ContainerStarted","Data":"80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c"} Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.503748 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.505528 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8d2f95cc-1c69-4731-90a9-a57b693b87ac","Type":"ContainerStarted","Data":"99d98b3fbf485162e166072bfecabe5915aaa2e14a16a4e3365edcb3213e19d7"} Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.516283 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dff579849-k2lss" podStartSLOduration=18.409156571 podStartE2EDuration="18.516261088s" podCreationTimestamp="2025-10-04 08:32:24 +0000 UTC" firstStartedPulling="2025-10-04 08:32:40.105048914 +0000 UTC m=+987.859317728" lastFinishedPulling="2025-10-04 08:32:40.212153431 +0000 UTC m=+987.966422245" observedRunningTime="2025-10-04 08:32:42.510918815 +0000 UTC m=+990.265187629" watchObservedRunningTime="2025-10-04 08:32:42.516261088 +0000 UTC m=+990.270529902" Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.583270 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bcc77c96f-6jtkx"] Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.589257 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bcc77c96f-6jtkx"] Oct 04 08:32:42 crc kubenswrapper[4969]: I1004 08:32:42.591092 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" podStartSLOduration=18.366974782 podStartE2EDuration="18.591076601s" podCreationTimestamp="2025-10-04 08:32:24 +0000 UTC" firstStartedPulling="2025-10-04 08:32:40.022043368 +0000 UTC m=+987.776312192" lastFinishedPulling="2025-10-04 08:32:40.246145197 +0000 UTC m=+988.000414011" observedRunningTime="2025-10-04 08:32:42.562878909 +0000 UTC m=+990.317147733" watchObservedRunningTime="2025-10-04 08:32:42.591076601 +0000 UTC m=+990.345345415" Oct 04 08:32:43 crc kubenswrapper[4969]: I1004 08:32:43.081878 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="790314fe-563f-4149-a058-74f7073c6895" path="/var/lib/kubelet/pods/790314fe-563f-4149-a058-74f7073c6895/volumes" Oct 04 08:32:43 crc kubenswrapper[4969]: I1004 08:32:43.082239 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="886c9045-f58f-419e-bf8c-90f5ae42cba0" path="/var/lib/kubelet/pods/886c9045-f58f-419e-bf8c-90f5ae42cba0/volumes" Oct 04 08:32:43 crc kubenswrapper[4969]: I1004 08:32:43.082704 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3d61042-65af-4739-a3d6-b820d577de2f" path="/var/lib/kubelet/pods/a3d61042-65af-4739-a3d6-b820d577de2f/volumes" Oct 04 08:32:49 crc kubenswrapper[4969]: I1004 08:32:49.667935 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:32:49 crc kubenswrapper[4969]: I1004 08:32:49.668442 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:32:49 crc kubenswrapper[4969]: I1004 08:32:49.668497 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:32:49 crc kubenswrapper[4969]: I1004 08:32:49.669294 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a0e7ef25f0eb92534c8c59058c5d643ea03dee7ad9ad7534679a9b23c47e914c"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:32:49 crc kubenswrapper[4969]: I1004 08:32:49.669363 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://a0e7ef25f0eb92534c8c59058c5d643ea03dee7ad9ad7534679a9b23c47e914c" gracePeriod=600 Oct 04 08:32:49 crc kubenswrapper[4969]: I1004 08:32:49.963731 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:50 crc kubenswrapper[4969]: I1004 08:32:50.303550 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:32:50 crc kubenswrapper[4969]: I1004 08:32:50.355554 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75d676bbb7-lhbhj"] Oct 04 08:32:50 crc kubenswrapper[4969]: I1004 08:32:50.579976 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"41d6e66e-6079-4278-a26c-04a299dd6d9d","Type":"ContainerStarted","Data":"eeba92592238157ec4b9f3008138b355e0f61650b1aff77a2cd5a22709ed9cb5"} Oct 04 08:32:50 crc kubenswrapper[4969]: I1004 08:32:50.580258 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 04 08:32:50 crc kubenswrapper[4969]: I1004 08:32:50.582393 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="a0e7ef25f0eb92534c8c59058c5d643ea03dee7ad9ad7534679a9b23c47e914c" exitCode=0 Oct 04 08:32:50 crc kubenswrapper[4969]: I1004 08:32:50.582577 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" podUID="b853223c-adde-4fa9-add8-2fe57bb240dd" containerName="dnsmasq-dns" containerID="cri-o://80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c" gracePeriod=10 Oct 04 08:32:50 crc kubenswrapper[4969]: I1004 08:32:50.582814 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"a0e7ef25f0eb92534c8c59058c5d643ea03dee7ad9ad7534679a9b23c47e914c"} Oct 04 08:32:50 crc kubenswrapper[4969]: I1004 08:32:50.582838 4969 scope.go:117] "RemoveContainer" containerID="47fb93c2519799da0fd36fa2c011f7f4604debf093aa2814af4cd13eb886ba6f" Oct 04 08:32:50 crc kubenswrapper[4969]: I1004 08:32:50.603110 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.396379065 podStartE2EDuration="21.603093303s" podCreationTimestamp="2025-10-04 08:32:29 +0000 UTC" firstStartedPulling="2025-10-04 08:32:41.092928157 +0000 UTC m=+988.847196971" lastFinishedPulling="2025-10-04 08:32:49.299642395 +0000 UTC m=+997.053911209" observedRunningTime="2025-10-04 08:32:50.596854478 +0000 UTC m=+998.351123292" watchObservedRunningTime="2025-10-04 08:32:50.603093303 +0000 UTC m=+998.357362117" Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.364644 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.460573 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-config\") pod \"b853223c-adde-4fa9-add8-2fe57bb240dd\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.460861 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-dns-svc\") pod \"b853223c-adde-4fa9-add8-2fe57bb240dd\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.460954 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dm84\" (UniqueName: \"kubernetes.io/projected/b853223c-adde-4fa9-add8-2fe57bb240dd-kube-api-access-5dm84\") pod \"b853223c-adde-4fa9-add8-2fe57bb240dd\" (UID: \"b853223c-adde-4fa9-add8-2fe57bb240dd\") " Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.590698 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88","Type":"ContainerStarted","Data":"8f346f3819c45dc7065ce80e9e05acfe8a7bf978c5f66292769291d8ff65c95a"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.592857 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"f952a2138e627189e2388732fb9b24c40ce1c3b15ba69934279b103aa755d5b4"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.594616 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8a3d9463-8792-487d-b4a5-1b61ab3310f6","Type":"ContainerStarted","Data":"64d8689e2870ff835a6491efc270ddd39b0bd15dd5ff08eda4b8c3ff6f3feb92"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.599629 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0e92558f-4f16-443e-b9a7-1c7f76013d2f","Type":"ContainerStarted","Data":"4d344506dabb32d7f6d8cab02f1a6d2d7772fd46272eab42fc4d4d7c1b3567ac"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.599864 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.602631 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"82fe815c-3f06-499f-9e3c-5a3b2f00d932","Type":"ContainerStarted","Data":"311d6b2437b1ce6ea31e31b4f619c52df671ffb1ca8dfde61c0e61ef42ea85a8"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.604507 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p9gfr" event={"ID":"3f6d0bb5-43f9-4c31-838b-f71927d68b08","Type":"ContainerStarted","Data":"949f9cefb76a6550fa1e3c6c692608f6ba479d2a5590487fe9aa46b280dcdd5a"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.607159 4969 generic.go:334] "Generic (PLEG): container finished" podID="b853223c-adde-4fa9-add8-2fe57bb240dd" containerID="80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c" exitCode=0 Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.607267 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.607435 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" event={"ID":"b853223c-adde-4fa9-add8-2fe57bb240dd","Type":"ContainerDied","Data":"80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.607484 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75d676bbb7-lhbhj" event={"ID":"b853223c-adde-4fa9-add8-2fe57bb240dd","Type":"ContainerDied","Data":"101be2359ef0f9e71951ea383ae5075252004773281afa7ff4f0519691d704da"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.607505 4969 scope.go:117] "RemoveContainer" containerID="80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c" Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.611412 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fms7j" event={"ID":"452b4c0a-fcb1-48ed-8c3a-87c34904a0a5","Type":"ContainerStarted","Data":"1a9a705d4d7f099cdcdd32e0af82552a4a6b673f99b54c0bdbcbdf531fa6f5f1"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.611617 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-fms7j" Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.615552 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8d2f95cc-1c69-4731-90a9-a57b693b87ac","Type":"ContainerStarted","Data":"128d8227456c338158c9235ffdc475e77c32ebee6aff2a31554c749599f8b316"} Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.701875 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.500037285 podStartE2EDuration="20.701859726s" podCreationTimestamp="2025-10-04 08:32:31 +0000 UTC" firstStartedPulling="2025-10-04 08:32:41.061113084 +0000 UTC m=+988.815381898" lastFinishedPulling="2025-10-04 08:32:50.262935525 +0000 UTC m=+998.017204339" observedRunningTime="2025-10-04 08:32:51.700319358 +0000 UTC m=+999.454588172" watchObservedRunningTime="2025-10-04 08:32:51.701859726 +0000 UTC m=+999.456128530" Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.722885 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-fms7j" podStartSLOduration=8.625834727 podStartE2EDuration="17.722865319s" podCreationTimestamp="2025-10-04 08:32:34 +0000 UTC" firstStartedPulling="2025-10-04 08:32:41.063483094 +0000 UTC m=+988.817751908" lastFinishedPulling="2025-10-04 08:32:50.160513686 +0000 UTC m=+997.914782500" observedRunningTime="2025-10-04 08:32:51.719388042 +0000 UTC m=+999.473656876" watchObservedRunningTime="2025-10-04 08:32:51.722865319 +0000 UTC m=+999.477134133" Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.746796 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b853223c-adde-4fa9-add8-2fe57bb240dd-kube-api-access-5dm84" (OuterVolumeSpecName: "kube-api-access-5dm84") pod "b853223c-adde-4fa9-add8-2fe57bb240dd" (UID: "b853223c-adde-4fa9-add8-2fe57bb240dd"). InnerVolumeSpecName "kube-api-access-5dm84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:32:51 crc kubenswrapper[4969]: I1004 08:32:51.767589 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dm84\" (UniqueName: \"kubernetes.io/projected/b853223c-adde-4fa9-add8-2fe57bb240dd-kube-api-access-5dm84\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.094745 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b853223c-adde-4fa9-add8-2fe57bb240dd" (UID: "b853223c-adde-4fa9-add8-2fe57bb240dd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.174544 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.369504 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-config" (OuterVolumeSpecName: "config") pod "b853223c-adde-4fa9-add8-2fe57bb240dd" (UID: "b853223c-adde-4fa9-add8-2fe57bb240dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.378322 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b853223c-adde-4fa9-add8-2fe57bb240dd-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.422962 4969 scope.go:117] "RemoveContainer" containerID="7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.445139 4969 scope.go:117] "RemoveContainer" containerID="80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c" Oct 04 08:32:52 crc kubenswrapper[4969]: E1004 08:32:52.450221 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c\": container with ID starting with 80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c not found: ID does not exist" containerID="80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.450274 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c"} err="failed to get container status \"80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c\": rpc error: code = NotFound desc = could not find container \"80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c\": container with ID starting with 80c06fde4df1469732abfbbbb1c865e140a65b0d474a55fe77886d99e93e2f3c not found: ID does not exist" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.450318 4969 scope.go:117] "RemoveContainer" containerID="7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4" Oct 04 08:32:52 crc kubenswrapper[4969]: E1004 08:32:52.450821 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4\": container with ID starting with 7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4 not found: ID does not exist" containerID="7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.450847 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4"} err="failed to get container status \"7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4\": rpc error: code = NotFound desc = could not find container \"7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4\": container with ID starting with 7862d4c1f5f74a094153cc361ca95f88b1e847f9f795cc58bdfa6a9f06040ec4 not found: ID does not exist" Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.553659 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75d676bbb7-lhbhj"] Oct 04 08:32:52 crc kubenswrapper[4969]: I1004 08:32:52.562163 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75d676bbb7-lhbhj"] Oct 04 08:32:53 crc kubenswrapper[4969]: I1004 08:32:53.075646 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b853223c-adde-4fa9-add8-2fe57bb240dd" path="/var/lib/kubelet/pods/b853223c-adde-4fa9-add8-2fe57bb240dd/volumes" Oct 04 08:32:53 crc kubenswrapper[4969]: I1004 08:32:53.634685 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6eaebec4-b71b-409d-b91e-1993605ce209","Type":"ContainerStarted","Data":"8b1f66b1e50805e2ec1fadbeadbbdb658b9a97373b0469c6ef10e190672e5c1c"} Oct 04 08:32:56 crc kubenswrapper[4969]: I1004 08:32:56.440901 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 04 08:32:56 crc kubenswrapper[4969]: I1004 08:32:56.669479 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1547b86a-ce65-4135-a8a5-957017c2271c","Type":"ContainerStarted","Data":"7293ab8d08343f9c68c67fb2c9a6ca5eae2fbf92ee2915c9b9c3d3620b5a04f8"} Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.070125 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-zssjt"] Oct 04 08:32:58 crc kubenswrapper[4969]: E1004 08:32:58.070843 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="886c9045-f58f-419e-bf8c-90f5ae42cba0" containerName="init" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.070855 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="886c9045-f58f-419e-bf8c-90f5ae42cba0" containerName="init" Oct 04 08:32:58 crc kubenswrapper[4969]: E1004 08:32:58.070875 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b853223c-adde-4fa9-add8-2fe57bb240dd" containerName="dnsmasq-dns" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.070881 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b853223c-adde-4fa9-add8-2fe57bb240dd" containerName="dnsmasq-dns" Oct 04 08:32:58 crc kubenswrapper[4969]: E1004 08:32:58.070906 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b853223c-adde-4fa9-add8-2fe57bb240dd" containerName="init" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.070913 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b853223c-adde-4fa9-add8-2fe57bb240dd" containerName="init" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.071053 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="886c9045-f58f-419e-bf8c-90f5ae42cba0" containerName="init" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.071062 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="b853223c-adde-4fa9-add8-2fe57bb240dd" containerName="dnsmasq-dns" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.071582 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.081412 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.099355 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-zssjt"] Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.192649 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59789dbbcf-k6fs8"] Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.194041 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.198655 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b09ae48-946d-458d-a119-194f28bd1081-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.198745 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b09ae48-946d-458d-a119-194f28bd1081-combined-ca-bundle\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.198788 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0b09ae48-946d-458d-a119-194f28bd1081-ovn-rundir\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.198819 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwhpp\" (UniqueName: \"kubernetes.io/projected/0b09ae48-946d-458d-a119-194f28bd1081-kube-api-access-cwhpp\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.198836 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0b09ae48-946d-458d-a119-194f28bd1081-ovs-rundir\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.198913 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b09ae48-946d-458d-a119-194f28bd1081-config\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.200092 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.201666 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59789dbbcf-k6fs8"] Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300347 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b09ae48-946d-458d-a119-194f28bd1081-config\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300447 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b09ae48-946d-458d-a119-194f28bd1081-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300518 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b09ae48-946d-458d-a119-194f28bd1081-combined-ca-bundle\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300554 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-config\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300656 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-ovsdbserver-nb\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300702 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0b09ae48-946d-458d-a119-194f28bd1081-ovn-rundir\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300731 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwhpp\" (UniqueName: \"kubernetes.io/projected/0b09ae48-946d-458d-a119-194f28bd1081-kube-api-access-cwhpp\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300748 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0b09ae48-946d-458d-a119-194f28bd1081-ovs-rundir\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300789 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-dns-svc\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.300812 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v52f\" (UniqueName: \"kubernetes.io/projected/5a106d12-744e-4a19-a341-f793785cca77-kube-api-access-6v52f\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.301519 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b09ae48-946d-458d-a119-194f28bd1081-config\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.301858 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0b09ae48-946d-458d-a119-194f28bd1081-ovn-rundir\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.302428 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0b09ae48-946d-458d-a119-194f28bd1081-ovs-rundir\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.310937 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b09ae48-946d-458d-a119-194f28bd1081-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.313039 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b09ae48-946d-458d-a119-194f28bd1081-combined-ca-bundle\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.320485 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwhpp\" (UniqueName: \"kubernetes.io/projected/0b09ae48-946d-458d-a119-194f28bd1081-kube-api-access-cwhpp\") pod \"ovn-controller-metrics-zssjt\" (UID: \"0b09ae48-946d-458d-a119-194f28bd1081\") " pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.396143 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-zssjt" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.401854 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-config\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.401898 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-ovsdbserver-nb\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.401943 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-dns-svc\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.401970 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v52f\" (UniqueName: \"kubernetes.io/projected/5a106d12-744e-4a19-a341-f793785cca77-kube-api-access-6v52f\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.403459 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-config\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.404020 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-ovsdbserver-nb\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.404020 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-dns-svc\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.428992 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v52f\" (UniqueName: \"kubernetes.io/projected/5a106d12-744e-4a19-a341-f793785cca77-kube-api-access-6v52f\") pod \"dnsmasq-dns-59789dbbcf-k6fs8\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.493304 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59789dbbcf-k6fs8"] Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.494196 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.522263 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b57f4fcb5-6rqk2"] Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.523602 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.528219 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.533859 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b57f4fcb5-6rqk2"] Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.609234 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-sb\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.609270 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-config\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.609350 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jm2l\" (UniqueName: \"kubernetes.io/projected/97f2cb01-2121-4a0c-a0d3-cce94e9088af-kube-api-access-9jm2l\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.609366 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-nb\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.609434 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-dns-svc\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.693030 4969 generic.go:334] "Generic (PLEG): container finished" podID="3f6d0bb5-43f9-4c31-838b-f71927d68b08" containerID="949f9cefb76a6550fa1e3c6c692608f6ba479d2a5590487fe9aa46b280dcdd5a" exitCode=0 Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.693102 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p9gfr" event={"ID":"3f6d0bb5-43f9-4c31-838b-f71927d68b08","Type":"ContainerDied","Data":"949f9cefb76a6550fa1e3c6c692608f6ba479d2a5590487fe9aa46b280dcdd5a"} Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.696891 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"28264a08-8168-419d-a20d-8f94a5c4ed77","Type":"ContainerStarted","Data":"6cfa4591c67cb36af29eb68362609a083c92363f9ed19aba47fa884f35f7da4c"} Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.699335 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerStarted","Data":"ae892b616ad06e64c04daa6d3f8d132e2326db8c507796a2df01a12da54d9b29"} Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.711059 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jm2l\" (UniqueName: \"kubernetes.io/projected/97f2cb01-2121-4a0c-a0d3-cce94e9088af-kube-api-access-9jm2l\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.711100 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-nb\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.711158 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-dns-svc\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.711195 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-sb\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.711211 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-config\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.712138 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-config\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.713169 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-nb\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.714110 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-dns-svc\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.714478 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-sb\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.750111 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jm2l\" (UniqueName: \"kubernetes.io/projected/97f2cb01-2121-4a0c-a0d3-cce94e9088af-kube-api-access-9jm2l\") pod \"dnsmasq-dns-7b57f4fcb5-6rqk2\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:32:58 crc kubenswrapper[4969]: I1004 08:32:58.842525 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.488723 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59789dbbcf-k6fs8"] Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.581926 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-zssjt"] Oct 04 08:33:00 crc kubenswrapper[4969]: W1004 08:33:00.597860 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b09ae48_946d_458d_a119_194f28bd1081.slice/crio-5c2db4068c1ffd70fcecfb0fae652ace96bdf87ba6066874fd0033760d8b4700 WatchSource:0}: Error finding container 5c2db4068c1ffd70fcecfb0fae652ace96bdf87ba6066874fd0033760d8b4700: Status 404 returned error can't find the container with id 5c2db4068c1ffd70fcecfb0fae652ace96bdf87ba6066874fd0033760d8b4700 Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.623131 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b57f4fcb5-6rqk2"] Oct 04 08:33:00 crc kubenswrapper[4969]: W1004 08:33:00.639672 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97f2cb01_2121_4a0c_a0d3_cce94e9088af.slice/crio-5e483326e8118a7c3ab681b1007449657a2d9e909eb5fc57d6f2f543e467cee4 WatchSource:0}: Error finding container 5e483326e8118a7c3ab681b1007449657a2d9e909eb5fc57d6f2f543e467cee4: Status 404 returned error can't find the container with id 5e483326e8118a7c3ab681b1007449657a2d9e909eb5fc57d6f2f543e467cee4 Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.722133 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" event={"ID":"5a106d12-744e-4a19-a341-f793785cca77","Type":"ContainerStarted","Data":"91338480bdc0565d32422892ed66e9caec0b30f4966ea866a720bb5ce92a6da2"} Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.723959 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-zssjt" event={"ID":"0b09ae48-946d-458d-a119-194f28bd1081","Type":"ContainerStarted","Data":"5c2db4068c1ffd70fcecfb0fae652ace96bdf87ba6066874fd0033760d8b4700"} Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.725586 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" event={"ID":"97f2cb01-2121-4a0c-a0d3-cce94e9088af","Type":"ContainerStarted","Data":"5e483326e8118a7c3ab681b1007449657a2d9e909eb5fc57d6f2f543e467cee4"} Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.727776 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8d2f95cc-1c69-4731-90a9-a57b693b87ac","Type":"ContainerStarted","Data":"2c5d20228522889a63ccb4a1c5cc4fd9bfce57f3e207b2bb17a4ddf6f25f20b9"} Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.730019 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d179b5df-13f6-4e3f-bbd5-d210ba6b2f88","Type":"ContainerStarted","Data":"8b1bf3e3a8dca4e428aa14cfbcdb7c7df0f0047f886d2951cd16ca1d695432bf"} Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.734225 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p9gfr" event={"ID":"3f6d0bb5-43f9-4c31-838b-f71927d68b08","Type":"ContainerStarted","Data":"0471335affc44aaccf8bd5258f52b288e1250d15e30b7bd801c2fa29ca066614"} Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.763370 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.074272546 podStartE2EDuration="26.763347944s" podCreationTimestamp="2025-10-04 08:32:34 +0000 UTC" firstStartedPulling="2025-10-04 08:32:41.381832308 +0000 UTC m=+989.136101122" lastFinishedPulling="2025-10-04 08:33:00.070907706 +0000 UTC m=+1007.825176520" observedRunningTime="2025-10-04 08:33:00.750681939 +0000 UTC m=+1008.504950753" watchObservedRunningTime="2025-10-04 08:33:00.763347944 +0000 UTC m=+1008.517616758" Oct 04 08:33:00 crc kubenswrapper[4969]: I1004 08:33:00.795186 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.864217806 podStartE2EDuration="23.795162286s" podCreationTimestamp="2025-10-04 08:32:37 +0000 UTC" firstStartedPulling="2025-10-04 08:32:41.15896643 +0000 UTC m=+988.913235244" lastFinishedPulling="2025-10-04 08:33:00.0899109 +0000 UTC m=+1007.844179724" observedRunningTime="2025-10-04 08:33:00.77402471 +0000 UTC m=+1008.528293544" watchObservedRunningTime="2025-10-04 08:33:00.795162286 +0000 UTC m=+1008.549431100" Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.749575 4969 generic.go:334] "Generic (PLEG): container finished" podID="5a106d12-744e-4a19-a341-f793785cca77" containerID="738deea023110971b9d8a3f27af7ec8051338197fc9a64f74edb104507bf80aa" exitCode=0 Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.749650 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" event={"ID":"5a106d12-744e-4a19-a341-f793785cca77","Type":"ContainerDied","Data":"738deea023110971b9d8a3f27af7ec8051338197fc9a64f74edb104507bf80aa"} Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.754238 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-zssjt" event={"ID":"0b09ae48-946d-458d-a119-194f28bd1081","Type":"ContainerStarted","Data":"5beefa1c3819d132e9fa791836d7a0a5d3bd07f677e1e7a9002931e1d39d3dd8"} Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.758507 4969 generic.go:334] "Generic (PLEG): container finished" podID="97f2cb01-2121-4a0c-a0d3-cce94e9088af" containerID="47927667e400f50d9b6bf761f3b68e5aea2c23edbd38dba655a5b2c865e6c4fe" exitCode=0 Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.758605 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" event={"ID":"97f2cb01-2121-4a0c-a0d3-cce94e9088af","Type":"ContainerDied","Data":"47927667e400f50d9b6bf761f3b68e5aea2c23edbd38dba655a5b2c865e6c4fe"} Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.768870 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p9gfr" event={"ID":"3f6d0bb5-43f9-4c31-838b-f71927d68b08","Type":"ContainerStarted","Data":"0702cc74c3254433f2dcbb790394fc87b72df77ee9c699537baa757ab14c3c1b"} Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.771087 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.771458 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.834339 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-zssjt" podStartSLOduration=3.8230673939999997 podStartE2EDuration="3.823067394s" podCreationTimestamp="2025-10-04 08:32:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:33:01.800905343 +0000 UTC m=+1009.555174187" watchObservedRunningTime="2025-10-04 08:33:01.823067394 +0000 UTC m=+1009.577336198" Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.904771 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-p9gfr" podStartSLOduration=20.606997778 podStartE2EDuration="27.904752168s" podCreationTimestamp="2025-10-04 08:32:34 +0000 UTC" firstStartedPulling="2025-10-04 08:32:42.163743014 +0000 UTC m=+989.918011828" lastFinishedPulling="2025-10-04 08:32:49.461497404 +0000 UTC m=+997.215766218" observedRunningTime="2025-10-04 08:33:01.874917416 +0000 UTC m=+1009.629186240" watchObservedRunningTime="2025-10-04 08:33:01.904752168 +0000 UTC m=+1009.659020972" Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.972981 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b57f4fcb5-6rqk2"] Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.995409 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79795ddc65-wj2fb"] Oct 04 08:33:01 crc kubenswrapper[4969]: I1004 08:33:01.996921 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.008871 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79795ddc65-wj2fb"] Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.025212 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 08:33:02 crc kubenswrapper[4969]: E1004 08:33:02.092388 4969 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 04 08:33:02 crc kubenswrapper[4969]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/97f2cb01-2121-4a0c-a0d3-cce94e9088af/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 04 08:33:02 crc kubenswrapper[4969]: > podSandboxID="5e483326e8118a7c3ab681b1007449657a2d9e909eb5fc57d6f2f543e467cee4" Oct 04 08:33:02 crc kubenswrapper[4969]: E1004 08:33:02.092960 4969 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 04 08:33:02 crc kubenswrapper[4969]: container &Container{Name:dnsmasq-dns,Image:38.102.83.83:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n7bh64fh67ch5c4h65bh587h67fh546h7bhc4h688h596h5c7h554h99h8h5dch586h7h5cbh686h55h64bh7dhdbhb6h575h65ch654h658h688h65bq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9jm2l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7b57f4fcb5-6rqk2_openstack(97f2cb01-2121-4a0c-a0d3-cce94e9088af): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/97f2cb01-2121-4a0c-a0d3-cce94e9088af/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 04 08:33:02 crc kubenswrapper[4969]: > logger="UnhandledError" Oct 04 08:33:02 crc kubenswrapper[4969]: E1004 08:33:02.094584 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/97f2cb01-2121-4a0c-a0d3-cce94e9088af/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" podUID="97f2cb01-2121-4a0c-a0d3-cce94e9088af" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.184205 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-sb\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.185486 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd6r8\" (UniqueName: \"kubernetes.io/projected/1de7855b-90f1-46ae-a85c-9e292224b87e-kube-api-access-kd6r8\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.185530 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-nb\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.185594 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-dns-svc\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.185678 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-config\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.287806 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-sb\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.287855 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd6r8\" (UniqueName: \"kubernetes.io/projected/1de7855b-90f1-46ae-a85c-9e292224b87e-kube-api-access-kd6r8\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.287876 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-nb\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.287909 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-dns-svc\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.287950 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-config\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.288731 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-config\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.289238 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-sb\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.290040 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-nb\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.290695 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-dns-svc\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.308566 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd6r8\" (UniqueName: \"kubernetes.io/projected/1de7855b-90f1-46ae-a85c-9e292224b87e-kube-api-access-kd6r8\") pod \"dnsmasq-dns-79795ddc65-wj2fb\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.316835 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.462295 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.593015 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-dns-svc\") pod \"5a106d12-744e-4a19-a341-f793785cca77\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.593074 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-ovsdbserver-nb\") pod \"5a106d12-744e-4a19-a341-f793785cca77\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.593141 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v52f\" (UniqueName: \"kubernetes.io/projected/5a106d12-744e-4a19-a341-f793785cca77-kube-api-access-6v52f\") pod \"5a106d12-744e-4a19-a341-f793785cca77\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.593280 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-config\") pod \"5a106d12-744e-4a19-a341-f793785cca77\" (UID: \"5a106d12-744e-4a19-a341-f793785cca77\") " Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.596284 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a106d12-744e-4a19-a341-f793785cca77-kube-api-access-6v52f" (OuterVolumeSpecName: "kube-api-access-6v52f") pod "5a106d12-744e-4a19-a341-f793785cca77" (UID: "5a106d12-744e-4a19-a341-f793785cca77"). InnerVolumeSpecName "kube-api-access-6v52f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.617542 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5a106d12-744e-4a19-a341-f793785cca77" (UID: "5a106d12-744e-4a19-a341-f793785cca77"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.617634 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5a106d12-744e-4a19-a341-f793785cca77" (UID: "5a106d12-744e-4a19-a341-f793785cca77"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.622430 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-config" (OuterVolumeSpecName: "config") pod "5a106d12-744e-4a19-a341-f793785cca77" (UID: "5a106d12-744e-4a19-a341-f793785cca77"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.671559 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.694989 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.695021 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.695031 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a106d12-744e-4a19-a341-f793785cca77-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.695044 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v52f\" (UniqueName: \"kubernetes.io/projected/5a106d12-744e-4a19-a341-f793785cca77-kube-api-access-6v52f\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.710993 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.713091 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.753705 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.776098 4969 generic.go:334] "Generic (PLEG): container finished" podID="8a3d9463-8792-487d-b4a5-1b61ab3310f6" containerID="64d8689e2870ff835a6491efc270ddd39b0bd15dd5ff08eda4b8c3ff6f3feb92" exitCode=0 Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.776179 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8a3d9463-8792-487d-b4a5-1b61ab3310f6","Type":"ContainerDied","Data":"64d8689e2870ff835a6491efc270ddd39b0bd15dd5ff08eda4b8c3ff6f3feb92"} Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.779358 4969 generic.go:334] "Generic (PLEG): container finished" podID="82fe815c-3f06-499f-9e3c-5a3b2f00d932" containerID="311d6b2437b1ce6ea31e31b4f619c52df671ffb1ca8dfde61c0e61ef42ea85a8" exitCode=0 Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.779413 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"82fe815c-3f06-499f-9e3c-5a3b2f00d932","Type":"ContainerDied","Data":"311d6b2437b1ce6ea31e31b4f619c52df671ffb1ca8dfde61c0e61ef42ea85a8"} Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.782383 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.784187 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59789dbbcf-k6fs8" event={"ID":"5a106d12-744e-4a19-a341-f793785cca77","Type":"ContainerDied","Data":"91338480bdc0565d32422892ed66e9caec0b30f4966ea866a720bb5ce92a6da2"} Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.784260 4969 scope.go:117] "RemoveContainer" containerID="738deea023110971b9d8a3f27af7ec8051338197fc9a64f74edb104507bf80aa" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.785324 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.785360 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 04 08:33:02 crc kubenswrapper[4969]: W1004 08:33:02.792663 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1de7855b_90f1_46ae_a85c_9e292224b87e.slice/crio-b5b8823e17beba8088dbdd8a7fb6ea4943737a8a988751f49dc5bb5862480c84 WatchSource:0}: Error finding container b5b8823e17beba8088dbdd8a7fb6ea4943737a8a988751f49dc5bb5862480c84: Status 404 returned error can't find the container with id b5b8823e17beba8088dbdd8a7fb6ea4943737a8a988751f49dc5bb5862480c84 Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.797175 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79795ddc65-wj2fb"] Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.832622 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.843493 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.884637 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59789dbbcf-k6fs8"] Oct 04 08:33:02 crc kubenswrapper[4969]: I1004 08:33:02.890655 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59789dbbcf-k6fs8"] Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.070062 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a106d12-744e-4a19-a341-f793785cca77" path="/var/lib/kubelet/pods/5a106d12-744e-4a19-a341-f793785cca77/volumes" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.133639 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 04 08:33:03 crc kubenswrapper[4969]: E1004 08:33:03.134003 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a106d12-744e-4a19-a341-f793785cca77" containerName="init" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.134023 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a106d12-744e-4a19-a341-f793785cca77" containerName="init" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.134224 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a106d12-744e-4a19-a341-f793785cca77" containerName="init" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.159575 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.161838 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.164728 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.164830 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.164916 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-swft5" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.194271 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.198280 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.199921 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.202402 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-sb9f8" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.202600 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.202709 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.202869 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.253973 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.258074 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.309406 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-dns-svc\") pod \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.309480 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-nb\") pod \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.309605 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-config\") pod \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.309663 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-sb\") pod \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.309690 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jm2l\" (UniqueName: \"kubernetes.io/projected/97f2cb01-2121-4a0c-a0d3-cce94e9088af-kube-api-access-9jm2l\") pod \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\" (UID: \"97f2cb01-2121-4a0c-a0d3-cce94e9088af\") " Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.309863 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/138cab2b-aabb-4736-a566-d9aee97753c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.309944 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gzrf\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-kube-api-access-7gzrf\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.309968 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.309984 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.310005 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/138cab2b-aabb-4736-a566-d9aee97753c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.310037 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-lock\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.310057 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/138cab2b-aabb-4736-a566-d9aee97753c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.310085 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/138cab2b-aabb-4736-a566-d9aee97753c0-scripts\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.310098 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/138cab2b-aabb-4736-a566-d9aee97753c0-config\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.310118 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/138cab2b-aabb-4736-a566-d9aee97753c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.310135 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p6nl\" (UniqueName: \"kubernetes.io/projected/138cab2b-aabb-4736-a566-d9aee97753c0-kube-api-access-4p6nl\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.310154 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-cache\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.333203 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97f2cb01-2121-4a0c-a0d3-cce94e9088af-kube-api-access-9jm2l" (OuterVolumeSpecName: "kube-api-access-9jm2l") pod "97f2cb01-2121-4a0c-a0d3-cce94e9088af" (UID: "97f2cb01-2121-4a0c-a0d3-cce94e9088af"). InnerVolumeSpecName "kube-api-access-9jm2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.354067 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-config" (OuterVolumeSpecName: "config") pod "97f2cb01-2121-4a0c-a0d3-cce94e9088af" (UID: "97f2cb01-2121-4a0c-a0d3-cce94e9088af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.354957 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "97f2cb01-2121-4a0c-a0d3-cce94e9088af" (UID: "97f2cb01-2121-4a0c-a0d3-cce94e9088af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.374188 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "97f2cb01-2121-4a0c-a0d3-cce94e9088af" (UID: "97f2cb01-2121-4a0c-a0d3-cce94e9088af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.385921 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "97f2cb01-2121-4a0c-a0d3-cce94e9088af" (UID: "97f2cb01-2121-4a0c-a0d3-cce94e9088af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.411876 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/138cab2b-aabb-4736-a566-d9aee97753c0-scripts\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.411920 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/138cab2b-aabb-4736-a566-d9aee97753c0-config\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.411954 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/138cab2b-aabb-4736-a566-d9aee97753c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.411979 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p6nl\" (UniqueName: \"kubernetes.io/projected/138cab2b-aabb-4736-a566-d9aee97753c0-kube-api-access-4p6nl\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412009 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-cache\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412040 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/138cab2b-aabb-4736-a566-d9aee97753c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412106 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gzrf\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-kube-api-access-7gzrf\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412138 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412179 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412209 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/138cab2b-aabb-4736-a566-d9aee97753c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412254 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-lock\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412284 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/138cab2b-aabb-4736-a566-d9aee97753c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412338 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412351 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412365 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jm2l\" (UniqueName: \"kubernetes.io/projected/97f2cb01-2121-4a0c-a0d3-cce94e9088af-kube-api-access-9jm2l\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412378 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.412389 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97f2cb01-2121-4a0c-a0d3-cce94e9088af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.413175 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.413308 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/138cab2b-aabb-4736-a566-d9aee97753c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: E1004 08:33:03.413665 4969 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 08:33:03 crc kubenswrapper[4969]: E1004 08:33:03.413697 4969 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 08:33:03 crc kubenswrapper[4969]: E1004 08:33:03.413801 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift podName:c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4 nodeName:}" failed. No retries permitted until 2025-10-04 08:33:03.913736294 +0000 UTC m=+1011.668005198 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift") pod "swift-storage-0" (UID: "c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4") : configmap "swift-ring-files" not found Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.414076 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-cache\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.414292 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-lock\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.414875 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/138cab2b-aabb-4736-a566-d9aee97753c0-scripts\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.416990 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/138cab2b-aabb-4736-a566-d9aee97753c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.423625 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/138cab2b-aabb-4736-a566-d9aee97753c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.425095 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/138cab2b-aabb-4736-a566-d9aee97753c0-config\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.431866 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/138cab2b-aabb-4736-a566-d9aee97753c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.434815 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gzrf\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-kube-api-access-7gzrf\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.442414 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p6nl\" (UniqueName: \"kubernetes.io/projected/138cab2b-aabb-4736-a566-d9aee97753c0-kube-api-access-4p6nl\") pod \"ovn-northd-0\" (UID: \"138cab2b-aabb-4736-a566-d9aee97753c0\") " pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.453027 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.462559 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-b822c"] Oct 04 08:33:03 crc kubenswrapper[4969]: E1004 08:33:03.463089 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f2cb01-2121-4a0c-a0d3-cce94e9088af" containerName="init" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.463113 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f2cb01-2121-4a0c-a0d3-cce94e9088af" containerName="init" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.463340 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="97f2cb01-2121-4a0c-a0d3-cce94e9088af" containerName="init" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.464106 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.466604 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.466866 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.473132 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-b822c"] Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.474457 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.514178 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-dispersionconf\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.514240 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-swiftconf\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.514271 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-scripts\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.514313 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfcw6\" (UniqueName: \"kubernetes.io/projected/67329c10-3dda-4a99-a964-cc8039752729-kube-api-access-lfcw6\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.514622 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67329c10-3dda-4a99-a964-cc8039752729-etc-swift\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.514679 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-combined-ca-bundle\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.514918 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-ring-data-devices\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.588012 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.615934 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67329c10-3dda-4a99-a964-cc8039752729-etc-swift\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.615975 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-combined-ca-bundle\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.616032 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-ring-data-devices\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.616069 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-dispersionconf\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.616091 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-swiftconf\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.616109 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-scripts\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.616134 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfcw6\" (UniqueName: \"kubernetes.io/projected/67329c10-3dda-4a99-a964-cc8039752729-kube-api-access-lfcw6\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.616494 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67329c10-3dda-4a99-a964-cc8039752729-etc-swift\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.619000 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-scripts\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.619472 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-dispersionconf\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.619862 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-ring-data-devices\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.619938 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-combined-ca-bundle\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.622847 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-swiftconf\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.634099 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfcw6\" (UniqueName: \"kubernetes.io/projected/67329c10-3dda-4a99-a964-cc8039752729-kube-api-access-lfcw6\") pod \"swift-ring-rebalance-b822c\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.788869 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.832831 4969 generic.go:334] "Generic (PLEG): container finished" podID="1de7855b-90f1-46ae-a85c-9e292224b87e" containerID="5cd4e588f18cfdc29b3af8abac292c60fe0dff0b4bc314b288658ec204b4747a" exitCode=0 Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.832917 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" event={"ID":"1de7855b-90f1-46ae-a85c-9e292224b87e","Type":"ContainerDied","Data":"5cd4e588f18cfdc29b3af8abac292c60fe0dff0b4bc314b288658ec204b4747a"} Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.832943 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" event={"ID":"1de7855b-90f1-46ae-a85c-9e292224b87e","Type":"ContainerStarted","Data":"b5b8823e17beba8088dbdd8a7fb6ea4943737a8a988751f49dc5bb5862480c84"} Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.855867 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" event={"ID":"97f2cb01-2121-4a0c-a0d3-cce94e9088af","Type":"ContainerDied","Data":"5e483326e8118a7c3ab681b1007449657a2d9e909eb5fc57d6f2f543e467cee4"} Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.855923 4969 scope.go:117] "RemoveContainer" containerID="47927667e400f50d9b6bf761f3b68e5aea2c23edbd38dba655a5b2c865e6c4fe" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.856047 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b57f4fcb5-6rqk2" Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.909974 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"82fe815c-3f06-499f-9e3c-5a3b2f00d932","Type":"ContainerStarted","Data":"f937f7481bb691113faea2be6d107d85bd9ec1672df736875318ee25c70aabeb"} Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.924893 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:03 crc kubenswrapper[4969]: E1004 08:33:03.925055 4969 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 08:33:03 crc kubenswrapper[4969]: E1004 08:33:03.925068 4969 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 08:33:03 crc kubenswrapper[4969]: E1004 08:33:03.925112 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift podName:c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4 nodeName:}" failed. No retries permitted until 2025-10-04 08:33:04.925097323 +0000 UTC m=+1012.679366137 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift") pod "swift-storage-0" (UID: "c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4") : configmap "swift-ring-files" not found Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.966851 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b57f4fcb5-6rqk2"] Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.966886 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8a3d9463-8792-487d-b4a5-1b61ab3310f6","Type":"ContainerStarted","Data":"195c0c5b2431cd244df3dbc62d7935c49f89c0be0c4fffc81d0001996d9f023d"} Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.989875 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b57f4fcb5-6rqk2"] Oct 04 08:33:03 crc kubenswrapper[4969]: I1004 08:33:03.990237 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=28.484103863 podStartE2EDuration="36.990216874s" podCreationTimestamp="2025-10-04 08:32:27 +0000 UTC" firstStartedPulling="2025-10-04 08:32:40.807332587 +0000 UTC m=+988.561601401" lastFinishedPulling="2025-10-04 08:32:49.313445588 +0000 UTC m=+997.067714412" observedRunningTime="2025-10-04 08:33:03.987979169 +0000 UTC m=+1011.742247983" watchObservedRunningTime="2025-10-04 08:33:03.990216874 +0000 UTC m=+1011.744485688" Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.024168 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=25.687203037 podStartE2EDuration="37.024127159s" podCreationTimestamp="2025-10-04 08:32:27 +0000 UTC" firstStartedPulling="2025-10-04 08:32:38.821684246 +0000 UTC m=+986.575953070" lastFinishedPulling="2025-10-04 08:32:50.158608368 +0000 UTC m=+997.912877192" observedRunningTime="2025-10-04 08:33:04.018355634 +0000 UTC m=+1011.772624458" watchObservedRunningTime="2025-10-04 08:33:04.024127159 +0000 UTC m=+1011.778395973" Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.075494 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.392322 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-b822c"] Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.948000 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:04 crc kubenswrapper[4969]: E1004 08:33:04.948227 4969 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 08:33:04 crc kubenswrapper[4969]: E1004 08:33:04.948244 4969 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 08:33:04 crc kubenswrapper[4969]: E1004 08:33:04.948295 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift podName:c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4 nodeName:}" failed. No retries permitted until 2025-10-04 08:33:06.948278305 +0000 UTC m=+1014.702547119 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift") pod "swift-storage-0" (UID: "c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4") : configmap "swift-ring-files" not found Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.977468 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"138cab2b-aabb-4736-a566-d9aee97753c0","Type":"ContainerStarted","Data":"955e4d65e499ae600b5728d7ebb313007df84a72c23828e6f92d85e2712c83bc"} Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.979468 4969 generic.go:334] "Generic (PLEG): container finished" podID="49069f68-580b-41f3-9504-6e8b722459f7" containerID="ae892b616ad06e64c04daa6d3f8d132e2326db8c507796a2df01a12da54d9b29" exitCode=0 Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.979551 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerDied","Data":"ae892b616ad06e64c04daa6d3f8d132e2326db8c507796a2df01a12da54d9b29"} Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.986665 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" event={"ID":"1de7855b-90f1-46ae-a85c-9e292224b87e","Type":"ContainerStarted","Data":"6149164f902be6d8b639a5f942ae0d60f89cff145bcca988a8150693ef37f0a0"} Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.986741 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:04 crc kubenswrapper[4969]: I1004 08:33:04.992284 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b822c" event={"ID":"67329c10-3dda-4a99-a964-cc8039752729","Type":"ContainerStarted","Data":"5eb27b489848835f318468bf6f26972e452e09653e40b19e566cbda1e4061f27"} Oct 04 08:33:05 crc kubenswrapper[4969]: I1004 08:33:05.026152 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" podStartSLOduration=4.026134322 podStartE2EDuration="4.026134322s" podCreationTimestamp="2025-10-04 08:33:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:33:05.019408465 +0000 UTC m=+1012.773677279" watchObservedRunningTime="2025-10-04 08:33:05.026134322 +0000 UTC m=+1012.780403136" Oct 04 08:33:05 crc kubenswrapper[4969]: I1004 08:33:05.079705 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97f2cb01-2121-4a0c-a0d3-cce94e9088af" path="/var/lib/kubelet/pods/97f2cb01-2121-4a0c-a0d3-cce94e9088af/volumes" Oct 04 08:33:06 crc kubenswrapper[4969]: I1004 08:33:06.010179 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"138cab2b-aabb-4736-a566-d9aee97753c0","Type":"ContainerStarted","Data":"a7d3ba6f4d3e9a59b0f6ef94e9621f418a32cced442eccf78d562c980f954ec5"} Oct 04 08:33:06 crc kubenswrapper[4969]: E1004 08:33:06.181725 4969 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.195:60032->38.102.83.195:43269: write tcp 38.102.83.195:60032->38.102.83.195:43269: write: broken pipe Oct 04 08:33:06 crc kubenswrapper[4969]: I1004 08:33:06.987036 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:06 crc kubenswrapper[4969]: E1004 08:33:06.987544 4969 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 08:33:06 crc kubenswrapper[4969]: E1004 08:33:06.987562 4969 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 08:33:06 crc kubenswrapper[4969]: E1004 08:33:06.987603 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift podName:c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4 nodeName:}" failed. No retries permitted until 2025-10-04 08:33:10.987590671 +0000 UTC m=+1018.741859485 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift") pod "swift-storage-0" (UID: "c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4") : configmap "swift-ring-files" not found Oct 04 08:33:08 crc kubenswrapper[4969]: I1004 08:33:08.030621 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"138cab2b-aabb-4736-a566-d9aee97753c0","Type":"ContainerStarted","Data":"74c991b66c65839918172df76c00e5c02bb9778bfd59d86b38409436d0a5fa0a"} Oct 04 08:33:08 crc kubenswrapper[4969]: I1004 08:33:08.030946 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 04 08:33:08 crc kubenswrapper[4969]: I1004 08:33:08.034249 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b822c" event={"ID":"67329c10-3dda-4a99-a964-cc8039752729","Type":"ContainerStarted","Data":"b750cec3f4ae9b252f3a907d29d728ed1e664b09582a64a3e8ef5bd48c57621b"} Oct 04 08:33:08 crc kubenswrapper[4969]: I1004 08:33:08.056091 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=4.411684348 podStartE2EDuration="5.05607084s" podCreationTimestamp="2025-10-04 08:33:03 +0000 UTC" firstStartedPulling="2025-10-04 08:33:04.124436265 +0000 UTC m=+1011.878705079" lastFinishedPulling="2025-10-04 08:33:04.768822757 +0000 UTC m=+1012.523091571" observedRunningTime="2025-10-04 08:33:08.046856391 +0000 UTC m=+1015.801125205" watchObservedRunningTime="2025-10-04 08:33:08.05607084 +0000 UTC m=+1015.810339654" Oct 04 08:33:08 crc kubenswrapper[4969]: I1004 08:33:08.063961 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-b822c" podStartSLOduration=2.655668704 podStartE2EDuration="5.063942296s" podCreationTimestamp="2025-10-04 08:33:03 +0000 UTC" firstStartedPulling="2025-10-04 08:33:04.408563408 +0000 UTC m=+1012.162832222" lastFinishedPulling="2025-10-04 08:33:06.81683699 +0000 UTC m=+1014.571105814" observedRunningTime="2025-10-04 08:33:08.06330975 +0000 UTC m=+1015.817578564" watchObservedRunningTime="2025-10-04 08:33:08.063942296 +0000 UTC m=+1015.818211110" Oct 04 08:33:08 crc kubenswrapper[4969]: I1004 08:33:08.435372 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 04 08:33:08 crc kubenswrapper[4969]: I1004 08:33:08.435722 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 04 08:33:09 crc kubenswrapper[4969]: I1004 08:33:09.352161 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 04 08:33:09 crc kubenswrapper[4969]: I1004 08:33:09.352498 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 04 08:33:09 crc kubenswrapper[4969]: I1004 08:33:09.429194 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 04 08:33:10 crc kubenswrapper[4969]: I1004 08:33:10.123110 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 04 08:33:10 crc kubenswrapper[4969]: I1004 08:33:10.512501 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 04 08:33:10 crc kubenswrapper[4969]: I1004 08:33:10.589247 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 04 08:33:11 crc kubenswrapper[4969]: I1004 08:33:11.061491 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:11 crc kubenswrapper[4969]: E1004 08:33:11.061668 4969 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 08:33:11 crc kubenswrapper[4969]: E1004 08:33:11.061988 4969 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 08:33:11 crc kubenswrapper[4969]: E1004 08:33:11.062061 4969 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift podName:c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4 nodeName:}" failed. No retries permitted until 2025-10-04 08:33:19.062024971 +0000 UTC m=+1026.816293785 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift") pod "swift-storage-0" (UID: "c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4") : configmap "swift-ring-files" not found Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.066410 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-xlvgm"] Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.069956 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-xlvgm" Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.099122 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-xlvgm"] Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.185000 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgvxj\" (UniqueName: \"kubernetes.io/projected/9b39fae4-bbfe-466b-a08a-52119cf4e4a6-kube-api-access-hgvxj\") pod \"watcher-db-create-xlvgm\" (UID: \"9b39fae4-bbfe-466b-a08a-52119cf4e4a6\") " pod="openstack/watcher-db-create-xlvgm" Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.286415 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgvxj\" (UniqueName: \"kubernetes.io/projected/9b39fae4-bbfe-466b-a08a-52119cf4e4a6-kube-api-access-hgvxj\") pod \"watcher-db-create-xlvgm\" (UID: \"9b39fae4-bbfe-466b-a08a-52119cf4e4a6\") " pod="openstack/watcher-db-create-xlvgm" Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.310484 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgvxj\" (UniqueName: \"kubernetes.io/projected/9b39fae4-bbfe-466b-a08a-52119cf4e4a6-kube-api-access-hgvxj\") pod \"watcher-db-create-xlvgm\" (UID: \"9b39fae4-bbfe-466b-a08a-52119cf4e4a6\") " pod="openstack/watcher-db-create-xlvgm" Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.318669 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.372657 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dff579849-k2lss"] Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.372954 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dff579849-k2lss" podUID="da1d60cc-fcef-4418-a7d1-2ea245c0a274" containerName="dnsmasq-dns" containerID="cri-o://29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f" gracePeriod=10 Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.399742 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-xlvgm" Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.887279 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-xlvgm"] Oct 04 08:33:12 crc kubenswrapper[4969]: I1004 08:33:12.895016 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:33:12 crc kubenswrapper[4969]: W1004 08:33:12.899641 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b39fae4_bbfe_466b_a08a_52119cf4e4a6.slice/crio-2b4ea4a92af24bcfeb58c3d0aa5a8de43d5a1eea4d053ece275fd1685567668f WatchSource:0}: Error finding container 2b4ea4a92af24bcfeb58c3d0aa5a8de43d5a1eea4d053ece275fd1685567668f: Status 404 returned error can't find the container with id 2b4ea4a92af24bcfeb58c3d0aa5a8de43d5a1eea4d053ece275fd1685567668f Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.008756 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-config\") pod \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.009000 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-dns-svc\") pod \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.009068 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59s5w\" (UniqueName: \"kubernetes.io/projected/da1d60cc-fcef-4418-a7d1-2ea245c0a274-kube-api-access-59s5w\") pod \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\" (UID: \"da1d60cc-fcef-4418-a7d1-2ea245c0a274\") " Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.014194 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da1d60cc-fcef-4418-a7d1-2ea245c0a274-kube-api-access-59s5w" (OuterVolumeSpecName: "kube-api-access-59s5w") pod "da1d60cc-fcef-4418-a7d1-2ea245c0a274" (UID: "da1d60cc-fcef-4418-a7d1-2ea245c0a274"). InnerVolumeSpecName "kube-api-access-59s5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.063334 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da1d60cc-fcef-4418-a7d1-2ea245c0a274" (UID: "da1d60cc-fcef-4418-a7d1-2ea245c0a274"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.071055 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-config" (OuterVolumeSpecName: "config") pod "da1d60cc-fcef-4418-a7d1-2ea245c0a274" (UID: "da1d60cc-fcef-4418-a7d1-2ea245c0a274"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.099211 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-xlvgm" event={"ID":"9b39fae4-bbfe-466b-a08a-52119cf4e4a6","Type":"ContainerStarted","Data":"30e093e2106f123922932ff85d99396ce44570943dee67da2c8233ecb3fad031"} Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.099661 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-xlvgm" event={"ID":"9b39fae4-bbfe-466b-a08a-52119cf4e4a6","Type":"ContainerStarted","Data":"2b4ea4a92af24bcfeb58c3d0aa5a8de43d5a1eea4d053ece275fd1685567668f"} Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.102750 4969 generic.go:334] "Generic (PLEG): container finished" podID="da1d60cc-fcef-4418-a7d1-2ea245c0a274" containerID="29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f" exitCode=0 Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.102806 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dff579849-k2lss" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.102837 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dff579849-k2lss" event={"ID":"da1d60cc-fcef-4418-a7d1-2ea245c0a274","Type":"ContainerDied","Data":"29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f"} Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.102864 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dff579849-k2lss" event={"ID":"da1d60cc-fcef-4418-a7d1-2ea245c0a274","Type":"ContainerDied","Data":"c9bca4ec707e534dff9efe812ce3d823b7d9db793d4e5c49efa9b5f3b2cae69d"} Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.102884 4969 scope.go:117] "RemoveContainer" containerID="29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.106127 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerStarted","Data":"0fc68e0c8ef687a06bb8efb6b04020362453bde79bf724ff5315525cd9b3fc7b"} Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.111949 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.112014 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da1d60cc-fcef-4418-a7d1-2ea245c0a274-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.112030 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59s5w\" (UniqueName: \"kubernetes.io/projected/da1d60cc-fcef-4418-a7d1-2ea245c0a274-kube-api-access-59s5w\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.125851 4969 scope.go:117] "RemoveContainer" containerID="ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.140262 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-create-xlvgm" podStartSLOduration=1.137398555 podStartE2EDuration="1.137398555s" podCreationTimestamp="2025-10-04 08:33:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:33:13.126795452 +0000 UTC m=+1020.881064266" watchObservedRunningTime="2025-10-04 08:33:13.137398555 +0000 UTC m=+1020.891667369" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.149437 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dff579849-k2lss"] Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.156082 4969 scope.go:117] "RemoveContainer" containerID="29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f" Oct 04 08:33:13 crc kubenswrapper[4969]: E1004 08:33:13.156762 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f\": container with ID starting with 29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f not found: ID does not exist" containerID="29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.156801 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f"} err="failed to get container status \"29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f\": rpc error: code = NotFound desc = could not find container \"29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f\": container with ID starting with 29467610d286e3c6e23a6766a3ce10dc9365a3ea23ec4495b4ee5d3acdbf001f not found: ID does not exist" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.156827 4969 scope.go:117] "RemoveContainer" containerID="ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80" Oct 04 08:33:13 crc kubenswrapper[4969]: E1004 08:33:13.158641 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80\": container with ID starting with ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80 not found: ID does not exist" containerID="ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.158672 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80"} err="failed to get container status \"ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80\": rpc error: code = NotFound desc = could not find container \"ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80\": container with ID starting with ba618b5db955f37ee07418cd2ae8e895db15e30db1fbc962489ed44ab39cca80 not found: ID does not exist" Oct 04 08:33:13 crc kubenswrapper[4969]: I1004 08:33:13.159829 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dff579849-k2lss"] Oct 04 08:33:14 crc kubenswrapper[4969]: I1004 08:33:14.124887 4969 generic.go:334] "Generic (PLEG): container finished" podID="9b39fae4-bbfe-466b-a08a-52119cf4e4a6" containerID="30e093e2106f123922932ff85d99396ce44570943dee67da2c8233ecb3fad031" exitCode=0 Oct 04 08:33:14 crc kubenswrapper[4969]: I1004 08:33:14.125323 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-xlvgm" event={"ID":"9b39fae4-bbfe-466b-a08a-52119cf4e4a6","Type":"ContainerDied","Data":"30e093e2106f123922932ff85d99396ce44570943dee67da2c8233ecb3fad031"} Oct 04 08:33:15 crc kubenswrapper[4969]: I1004 08:33:15.075705 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da1d60cc-fcef-4418-a7d1-2ea245c0a274" path="/var/lib/kubelet/pods/da1d60cc-fcef-4418-a7d1-2ea245c0a274/volumes" Oct 04 08:33:15 crc kubenswrapper[4969]: I1004 08:33:15.538393 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-xlvgm" Oct 04 08:33:15 crc kubenswrapper[4969]: I1004 08:33:15.670110 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgvxj\" (UniqueName: \"kubernetes.io/projected/9b39fae4-bbfe-466b-a08a-52119cf4e4a6-kube-api-access-hgvxj\") pod \"9b39fae4-bbfe-466b-a08a-52119cf4e4a6\" (UID: \"9b39fae4-bbfe-466b-a08a-52119cf4e4a6\") " Oct 04 08:33:15 crc kubenswrapper[4969]: I1004 08:33:15.677273 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b39fae4-bbfe-466b-a08a-52119cf4e4a6-kube-api-access-hgvxj" (OuterVolumeSpecName: "kube-api-access-hgvxj") pod "9b39fae4-bbfe-466b-a08a-52119cf4e4a6" (UID: "9b39fae4-bbfe-466b-a08a-52119cf4e4a6"). InnerVolumeSpecName "kube-api-access-hgvxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:15 crc kubenswrapper[4969]: I1004 08:33:15.773099 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgvxj\" (UniqueName: \"kubernetes.io/projected/9b39fae4-bbfe-466b-a08a-52119cf4e4a6-kube-api-access-hgvxj\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.157270 4969 generic.go:334] "Generic (PLEG): container finished" podID="67329c10-3dda-4a99-a964-cc8039752729" containerID="b750cec3f4ae9b252f3a907d29d728ed1e664b09582a64a3e8ef5bd48c57621b" exitCode=0 Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.157351 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b822c" event={"ID":"67329c10-3dda-4a99-a964-cc8039752729","Type":"ContainerDied","Data":"b750cec3f4ae9b252f3a907d29d728ed1e664b09582a64a3e8ef5bd48c57621b"} Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.160013 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-xlvgm" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.160009 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-xlvgm" event={"ID":"9b39fae4-bbfe-466b-a08a-52119cf4e4a6","Type":"ContainerDied","Data":"2b4ea4a92af24bcfeb58c3d0aa5a8de43d5a1eea4d053ece275fd1685567668f"} Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.160100 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b4ea4a92af24bcfeb58c3d0aa5a8de43d5a1eea4d053ece275fd1685567668f" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.163054 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerStarted","Data":"488038e962c9c490bfc090e32286e992e959221611857e4f0b3009eb80ffb8ac"} Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.612387 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-7qcfn"] Oct 04 08:33:16 crc kubenswrapper[4969]: E1004 08:33:16.612940 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b39fae4-bbfe-466b-a08a-52119cf4e4a6" containerName="mariadb-database-create" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.612965 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b39fae4-bbfe-466b-a08a-52119cf4e4a6" containerName="mariadb-database-create" Oct 04 08:33:16 crc kubenswrapper[4969]: E1004 08:33:16.612994 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da1d60cc-fcef-4418-a7d1-2ea245c0a274" containerName="dnsmasq-dns" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.613007 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="da1d60cc-fcef-4418-a7d1-2ea245c0a274" containerName="dnsmasq-dns" Oct 04 08:33:16 crc kubenswrapper[4969]: E1004 08:33:16.613050 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da1d60cc-fcef-4418-a7d1-2ea245c0a274" containerName="init" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.613065 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="da1d60cc-fcef-4418-a7d1-2ea245c0a274" containerName="init" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.613365 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b39fae4-bbfe-466b-a08a-52119cf4e4a6" containerName="mariadb-database-create" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.613411 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="da1d60cc-fcef-4418-a7d1-2ea245c0a274" containerName="dnsmasq-dns" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.614453 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7qcfn" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.622083 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-7qcfn"] Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.690735 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hdtt\" (UniqueName: \"kubernetes.io/projected/2f9ae792-20d2-4617-9dd4-3b8ad58b12db-kube-api-access-5hdtt\") pod \"glance-db-create-7qcfn\" (UID: \"2f9ae792-20d2-4617-9dd4-3b8ad58b12db\") " pod="openstack/glance-db-create-7qcfn" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.792950 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hdtt\" (UniqueName: \"kubernetes.io/projected/2f9ae792-20d2-4617-9dd4-3b8ad58b12db-kube-api-access-5hdtt\") pod \"glance-db-create-7qcfn\" (UID: \"2f9ae792-20d2-4617-9dd4-3b8ad58b12db\") " pod="openstack/glance-db-create-7qcfn" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.840772 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hdtt\" (UniqueName: \"kubernetes.io/projected/2f9ae792-20d2-4617-9dd4-3b8ad58b12db-kube-api-access-5hdtt\") pod \"glance-db-create-7qcfn\" (UID: \"2f9ae792-20d2-4617-9dd4-3b8ad58b12db\") " pod="openstack/glance-db-create-7qcfn" Oct 04 08:33:16 crc kubenswrapper[4969]: I1004 08:33:16.935182 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7qcfn" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.242081 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-7qcfn"] Oct 04 08:33:17 crc kubenswrapper[4969]: W1004 08:33:17.257718 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f9ae792_20d2_4617_9dd4_3b8ad58b12db.slice/crio-3c08f72f259a6990e675b788d6b9281fee50ab41723eeb25225da171b5a0d440 WatchSource:0}: Error finding container 3c08f72f259a6990e675b788d6b9281fee50ab41723eeb25225da171b5a0d440: Status 404 returned error can't find the container with id 3c08f72f259a6990e675b788d6b9281fee50ab41723eeb25225da171b5a0d440 Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.484353 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.607769 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-swiftconf\") pod \"67329c10-3dda-4a99-a964-cc8039752729\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.607819 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67329c10-3dda-4a99-a964-cc8039752729-etc-swift\") pod \"67329c10-3dda-4a99-a964-cc8039752729\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.607859 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-ring-data-devices\") pod \"67329c10-3dda-4a99-a964-cc8039752729\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.607969 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-combined-ca-bundle\") pod \"67329c10-3dda-4a99-a964-cc8039752729\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.608067 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfcw6\" (UniqueName: \"kubernetes.io/projected/67329c10-3dda-4a99-a964-cc8039752729-kube-api-access-lfcw6\") pod \"67329c10-3dda-4a99-a964-cc8039752729\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.608091 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-dispersionconf\") pod \"67329c10-3dda-4a99-a964-cc8039752729\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.608107 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-scripts\") pod \"67329c10-3dda-4a99-a964-cc8039752729\" (UID: \"67329c10-3dda-4a99-a964-cc8039752729\") " Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.615247 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "67329c10-3dda-4a99-a964-cc8039752729" (UID: "67329c10-3dda-4a99-a964-cc8039752729"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.615409 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67329c10-3dda-4a99-a964-cc8039752729-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "67329c10-3dda-4a99-a964-cc8039752729" (UID: "67329c10-3dda-4a99-a964-cc8039752729"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.617363 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67329c10-3dda-4a99-a964-cc8039752729-kube-api-access-lfcw6" (OuterVolumeSpecName: "kube-api-access-lfcw6") pod "67329c10-3dda-4a99-a964-cc8039752729" (UID: "67329c10-3dda-4a99-a964-cc8039752729"). InnerVolumeSpecName "kube-api-access-lfcw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.623706 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "67329c10-3dda-4a99-a964-cc8039752729" (UID: "67329c10-3dda-4a99-a964-cc8039752729"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.630278 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-scripts" (OuterVolumeSpecName: "scripts") pod "67329c10-3dda-4a99-a964-cc8039752729" (UID: "67329c10-3dda-4a99-a964-cc8039752729"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.633094 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "67329c10-3dda-4a99-a964-cc8039752729" (UID: "67329c10-3dda-4a99-a964-cc8039752729"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.634759 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67329c10-3dda-4a99-a964-cc8039752729" (UID: "67329c10-3dda-4a99-a964-cc8039752729"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.709412 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.709455 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfcw6\" (UniqueName: \"kubernetes.io/projected/67329c10-3dda-4a99-a964-cc8039752729-kube-api-access-lfcw6\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.709467 4969 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.709477 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.709485 4969 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67329c10-3dda-4a99-a964-cc8039752729-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.709495 4969 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67329c10-3dda-4a99-a964-cc8039752729-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:17 crc kubenswrapper[4969]: I1004 08:33:17.709503 4969 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67329c10-3dda-4a99-a964-cc8039752729-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:18 crc kubenswrapper[4969]: I1004 08:33:18.210557 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7qcfn" event={"ID":"2f9ae792-20d2-4617-9dd4-3b8ad58b12db","Type":"ContainerStarted","Data":"3c08f72f259a6990e675b788d6b9281fee50ab41723eeb25225da171b5a0d440"} Oct 04 08:33:18 crc kubenswrapper[4969]: I1004 08:33:18.214220 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b822c" event={"ID":"67329c10-3dda-4a99-a964-cc8039752729","Type":"ContainerDied","Data":"5eb27b489848835f318468bf6f26972e452e09653e40b19e566cbda1e4061f27"} Oct 04 08:33:18 crc kubenswrapper[4969]: I1004 08:33:18.214284 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eb27b489848835f318468bf6f26972e452e09653e40b19e566cbda1e4061f27" Oct 04 08:33:18 crc kubenswrapper[4969]: I1004 08:33:18.214364 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b822c" Oct 04 08:33:18 crc kubenswrapper[4969]: I1004 08:33:18.691466 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.132848 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.139295 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4-etc-swift\") pod \"swift-storage-0\" (UID: \"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4\") " pod="openstack/swift-storage-0" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.182570 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.232945 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7qcfn" event={"ID":"2f9ae792-20d2-4617-9dd4-3b8ad58b12db","Type":"ContainerStarted","Data":"739922c6c43eef8820ea06f48bc977eb83476eeeeca4745b06309ef31bfbda92"} Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.568559 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-5zcjl"] Oct 04 08:33:19 crc kubenswrapper[4969]: E1004 08:33:19.569248 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67329c10-3dda-4a99-a964-cc8039752729" containerName="swift-ring-rebalance" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.569275 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="67329c10-3dda-4a99-a964-cc8039752729" containerName="swift-ring-rebalance" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.569543 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="67329c10-3dda-4a99-a964-cc8039752729" containerName="swift-ring-rebalance" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.570201 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-5zcjl" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.586845 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-5zcjl"] Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.638891 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvdcf\" (UniqueName: \"kubernetes.io/projected/d23e05af-c71b-4c8d-ac33-0a0aafbf3e32-kube-api-access-bvdcf\") pod \"keystone-db-create-5zcjl\" (UID: \"d23e05af-c71b-4c8d-ac33-0a0aafbf3e32\") " pod="openstack/keystone-db-create-5zcjl" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.726176 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.746249 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvdcf\" (UniqueName: \"kubernetes.io/projected/d23e05af-c71b-4c8d-ac33-0a0aafbf3e32-kube-api-access-bvdcf\") pod \"keystone-db-create-5zcjl\" (UID: \"d23e05af-c71b-4c8d-ac33-0a0aafbf3e32\") " pod="openstack/keystone-db-create-5zcjl" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.765902 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvdcf\" (UniqueName: \"kubernetes.io/projected/d23e05af-c71b-4c8d-ac33-0a0aafbf3e32-kube-api-access-bvdcf\") pod \"keystone-db-create-5zcjl\" (UID: \"d23e05af-c71b-4c8d-ac33-0a0aafbf3e32\") " pod="openstack/keystone-db-create-5zcjl" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.867352 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-cg6hf"] Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.870722 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cg6hf" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.876074 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cg6hf"] Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.919927 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-5zcjl" Oct 04 08:33:19 crc kubenswrapper[4969]: I1004 08:33:19.963903 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9n9h\" (UniqueName: \"kubernetes.io/projected/42d6b82d-a2fa-463e-8075-04bf086f145b-kube-api-access-s9n9h\") pod \"placement-db-create-cg6hf\" (UID: \"42d6b82d-a2fa-463e-8075-04bf086f145b\") " pod="openstack/placement-db-create-cg6hf" Oct 04 08:33:20 crc kubenswrapper[4969]: I1004 08:33:20.071861 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9n9h\" (UniqueName: \"kubernetes.io/projected/42d6b82d-a2fa-463e-8075-04bf086f145b-kube-api-access-s9n9h\") pod \"placement-db-create-cg6hf\" (UID: \"42d6b82d-a2fa-463e-8075-04bf086f145b\") " pod="openstack/placement-db-create-cg6hf" Oct 04 08:33:20 crc kubenswrapper[4969]: I1004 08:33:20.095446 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9n9h\" (UniqueName: \"kubernetes.io/projected/42d6b82d-a2fa-463e-8075-04bf086f145b-kube-api-access-s9n9h\") pod \"placement-db-create-cg6hf\" (UID: \"42d6b82d-a2fa-463e-8075-04bf086f145b\") " pod="openstack/placement-db-create-cg6hf" Oct 04 08:33:20 crc kubenswrapper[4969]: I1004 08:33:20.194297 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cg6hf" Oct 04 08:33:20 crc kubenswrapper[4969]: I1004 08:33:20.241325 4969 generic.go:334] "Generic (PLEG): container finished" podID="2f9ae792-20d2-4617-9dd4-3b8ad58b12db" containerID="739922c6c43eef8820ea06f48bc977eb83476eeeeca4745b06309ef31bfbda92" exitCode=0 Oct 04 08:33:20 crc kubenswrapper[4969]: I1004 08:33:20.241373 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7qcfn" event={"ID":"2f9ae792-20d2-4617-9dd4-3b8ad58b12db","Type":"ContainerDied","Data":"739922c6c43eef8820ea06f48bc977eb83476eeeeca4745b06309ef31bfbda92"} Oct 04 08:33:20 crc kubenswrapper[4969]: I1004 08:33:20.894122 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cg6hf"] Oct 04 08:33:20 crc kubenswrapper[4969]: I1004 08:33:20.950195 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-5zcjl"] Oct 04 08:33:20 crc kubenswrapper[4969]: W1004 08:33:20.982484 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42d6b82d_a2fa_463e_8075_04bf086f145b.slice/crio-65adb97bf57d4ea98b0930f72e8c17c4a1640e480b76460730f3fbf83e6b59b1 WatchSource:0}: Error finding container 65adb97bf57d4ea98b0930f72e8c17c4a1640e480b76460730f3fbf83e6b59b1: Status 404 returned error can't find the container with id 65adb97bf57d4ea98b0930f72e8c17c4a1640e480b76460730f3fbf83e6b59b1 Oct 04 08:33:20 crc kubenswrapper[4969]: W1004 08:33:20.985482 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd23e05af_c71b_4c8d_ac33_0a0aafbf3e32.slice/crio-20b8c0fe6f807eb1824f5d72965832f3316b846b286e699ae7bc3d4c1ec05d60 WatchSource:0}: Error finding container 20b8c0fe6f807eb1824f5d72965832f3316b846b286e699ae7bc3d4c1ec05d60: Status 404 returned error can't find the container with id 20b8c0fe6f807eb1824f5d72965832f3316b846b286e699ae7bc3d4c1ec05d60 Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.253637 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerStarted","Data":"2edabe18cb5eb64471d46ba83a6f332ca980c9481e667db15cafd0f1131e0a64"} Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.256300 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-5zcjl" event={"ID":"d23e05af-c71b-4c8d-ac33-0a0aafbf3e32","Type":"ContainerStarted","Data":"5a4f7d989a81b1e07ece8912930b2c83ea8cab66f8d80948f68b6950842f204e"} Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.256348 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-5zcjl" event={"ID":"d23e05af-c71b-4c8d-ac33-0a0aafbf3e32","Type":"ContainerStarted","Data":"20b8c0fe6f807eb1824f5d72965832f3316b846b286e699ae7bc3d4c1ec05d60"} Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.261055 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cg6hf" event={"ID":"42d6b82d-a2fa-463e-8075-04bf086f145b","Type":"ContainerStarted","Data":"3902102d442378626146542a829f8ccfc51b5df600b8bb5eb0ba844cf03fb8b1"} Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.261092 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cg6hf" event={"ID":"42d6b82d-a2fa-463e-8075-04bf086f145b","Type":"ContainerStarted","Data":"65adb97bf57d4ea98b0930f72e8c17c4a1640e480b76460730f3fbf83e6b59b1"} Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.264955 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"aae65612429ebb1a4468e72e1067f221afe493a0ec8a3f640c116b69fa722b62"} Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.264995 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"10c6277a5298b7e11794a140699264d408b00d21b5678d81d5fc047a23b49f8d"} Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.290571 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=10.871742195 podStartE2EDuration="50.29055139s" podCreationTimestamp="2025-10-04 08:32:31 +0000 UTC" firstStartedPulling="2025-10-04 08:32:41.063572076 +0000 UTC m=+988.817840890" lastFinishedPulling="2025-10-04 08:33:20.482381271 +0000 UTC m=+1028.236650085" observedRunningTime="2025-10-04 08:33:21.28411477 +0000 UTC m=+1029.038383594" watchObservedRunningTime="2025-10-04 08:33:21.29055139 +0000 UTC m=+1029.044820204" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.308946 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-5zcjl" podStartSLOduration=2.308919768 podStartE2EDuration="2.308919768s" podCreationTimestamp="2025-10-04 08:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:33:21.306625341 +0000 UTC m=+1029.060894155" watchObservedRunningTime="2025-10-04 08:33:21.308919768 +0000 UTC m=+1029.063188592" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.645178 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7qcfn" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.800146 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hdtt\" (UniqueName: \"kubernetes.io/projected/2f9ae792-20d2-4617-9dd4-3b8ad58b12db-kube-api-access-5hdtt\") pod \"2f9ae792-20d2-4617-9dd4-3b8ad58b12db\" (UID: \"2f9ae792-20d2-4617-9dd4-3b8ad58b12db\") " Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.808090 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f9ae792-20d2-4617-9dd4-3b8ad58b12db-kube-api-access-5hdtt" (OuterVolumeSpecName: "kube-api-access-5hdtt") pod "2f9ae792-20d2-4617-9dd4-3b8ad58b12db" (UID: "2f9ae792-20d2-4617-9dd4-3b8ad58b12db"). InnerVolumeSpecName "kube-api-access-5hdtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.902662 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hdtt\" (UniqueName: \"kubernetes.io/projected/2f9ae792-20d2-4617-9dd4-3b8ad58b12db-kube-api-access-5hdtt\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.975659 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-5c31-account-create-cqhsm"] Oct 04 08:33:21 crc kubenswrapper[4969]: E1004 08:33:21.975988 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9ae792-20d2-4617-9dd4-3b8ad58b12db" containerName="mariadb-database-create" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.976006 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9ae792-20d2-4617-9dd4-3b8ad58b12db" containerName="mariadb-database-create" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.976218 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f9ae792-20d2-4617-9dd4-3b8ad58b12db" containerName="mariadb-database-create" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.976782 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-5c31-account-create-cqhsm" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.979682 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Oct 04 08:33:21 crc kubenswrapper[4969]: I1004 08:33:21.985375 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-5c31-account-create-cqhsm"] Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.106736 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbvwt\" (UniqueName: \"kubernetes.io/projected/472bf752-8d20-4b60-82f8-71c73e4f8843-kube-api-access-dbvwt\") pod \"watcher-5c31-account-create-cqhsm\" (UID: \"472bf752-8d20-4b60-82f8-71c73e4f8843\") " pod="openstack/watcher-5c31-account-create-cqhsm" Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.208887 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbvwt\" (UniqueName: \"kubernetes.io/projected/472bf752-8d20-4b60-82f8-71c73e4f8843-kube-api-access-dbvwt\") pod \"watcher-5c31-account-create-cqhsm\" (UID: \"472bf752-8d20-4b60-82f8-71c73e4f8843\") " pod="openstack/watcher-5c31-account-create-cqhsm" Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.226760 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbvwt\" (UniqueName: \"kubernetes.io/projected/472bf752-8d20-4b60-82f8-71c73e4f8843-kube-api-access-dbvwt\") pod \"watcher-5c31-account-create-cqhsm\" (UID: \"472bf752-8d20-4b60-82f8-71c73e4f8843\") " pod="openstack/watcher-5c31-account-create-cqhsm" Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.277257 4969 generic.go:334] "Generic (PLEG): container finished" podID="d23e05af-c71b-4c8d-ac33-0a0aafbf3e32" containerID="5a4f7d989a81b1e07ece8912930b2c83ea8cab66f8d80948f68b6950842f204e" exitCode=0 Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.277807 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-5zcjl" event={"ID":"d23e05af-c71b-4c8d-ac33-0a0aafbf3e32","Type":"ContainerDied","Data":"5a4f7d989a81b1e07ece8912930b2c83ea8cab66f8d80948f68b6950842f204e"} Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.279592 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7qcfn" event={"ID":"2f9ae792-20d2-4617-9dd4-3b8ad58b12db","Type":"ContainerDied","Data":"3c08f72f259a6990e675b788d6b9281fee50ab41723eeb25225da171b5a0d440"} Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.279620 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c08f72f259a6990e675b788d6b9281fee50ab41723eeb25225da171b5a0d440" Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.279669 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7qcfn" Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.283396 4969 generic.go:334] "Generic (PLEG): container finished" podID="42d6b82d-a2fa-463e-8075-04bf086f145b" containerID="3902102d442378626146542a829f8ccfc51b5df600b8bb5eb0ba844cf03fb8b1" exitCode=0 Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.283531 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cg6hf" event={"ID":"42d6b82d-a2fa-463e-8075-04bf086f145b","Type":"ContainerDied","Data":"3902102d442378626146542a829f8ccfc51b5df600b8bb5eb0ba844cf03fb8b1"} Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.292684 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"1dc8f18f0b6664f35a516d8d73bf7a1e2896d3459c676c965b18a9f196510368"} Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.292811 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"705aa34ee6706a13aaae7179e44fc7e0f54ed67b84a9a7f49cdb000af55bba90"} Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.292970 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-5c31-account-create-cqhsm" Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.618853 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cg6hf" Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.716460 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9n9h\" (UniqueName: \"kubernetes.io/projected/42d6b82d-a2fa-463e-8075-04bf086f145b-kube-api-access-s9n9h\") pod \"42d6b82d-a2fa-463e-8075-04bf086f145b\" (UID: \"42d6b82d-a2fa-463e-8075-04bf086f145b\") " Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.720741 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d6b82d-a2fa-463e-8075-04bf086f145b-kube-api-access-s9n9h" (OuterVolumeSpecName: "kube-api-access-s9n9h") pod "42d6b82d-a2fa-463e-8075-04bf086f145b" (UID: "42d6b82d-a2fa-463e-8075-04bf086f145b"). InnerVolumeSpecName "kube-api-access-s9n9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.776755 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-5c31-account-create-cqhsm"] Oct 04 08:33:22 crc kubenswrapper[4969]: I1004 08:33:22.818252 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9n9h\" (UniqueName: \"kubernetes.io/projected/42d6b82d-a2fa-463e-8075-04bf086f145b-kube-api-access-s9n9h\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:22 crc kubenswrapper[4969]: W1004 08:33:22.820264 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod472bf752_8d20_4b60_82f8_71c73e4f8843.slice/crio-f88489c00e2c6768c7078489a1b70eb08300074207a6647cbf242ce2a82db1b5 WatchSource:0}: Error finding container f88489c00e2c6768c7078489a1b70eb08300074207a6647cbf242ce2a82db1b5: Status 404 returned error can't find the container with id f88489c00e2c6768c7078489a1b70eb08300074207a6647cbf242ce2a82db1b5 Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.192008 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.305280 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"50bdb4b23b4a30c9fe23c1218316092b2e35b034c3e75efa394702ec3f94b1ef"} Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.305349 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"2aae2b0844abe25d39e9ae23b5fc08f7f52a912067d379145868570861361a9b"} Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.305360 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"1a63aa76554fb06500a36ee383ea6bf23bf8e01c2deb5aaa7d52b54131711a15"} Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.306870 4969 generic.go:334] "Generic (PLEG): container finished" podID="472bf752-8d20-4b60-82f8-71c73e4f8843" containerID="c8da89e1e2530e1a8b0c314ecb957c582482a1301969f958bac691bad4b3aa85" exitCode=0 Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.306929 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-5c31-account-create-cqhsm" event={"ID":"472bf752-8d20-4b60-82f8-71c73e4f8843","Type":"ContainerDied","Data":"c8da89e1e2530e1a8b0c314ecb957c582482a1301969f958bac691bad4b3aa85"} Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.306947 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-5c31-account-create-cqhsm" event={"ID":"472bf752-8d20-4b60-82f8-71c73e4f8843","Type":"ContainerStarted","Data":"f88489c00e2c6768c7078489a1b70eb08300074207a6647cbf242ce2a82db1b5"} Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.309953 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cg6hf" Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.309752 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cg6hf" event={"ID":"42d6b82d-a2fa-463e-8075-04bf086f145b","Type":"ContainerDied","Data":"65adb97bf57d4ea98b0930f72e8c17c4a1640e480b76460730f3fbf83e6b59b1"} Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.310168 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65adb97bf57d4ea98b0930f72e8c17c4a1640e480b76460730f3fbf83e6b59b1" Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.729539 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-5zcjl" Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.832403 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvdcf\" (UniqueName: \"kubernetes.io/projected/d23e05af-c71b-4c8d-ac33-0a0aafbf3e32-kube-api-access-bvdcf\") pod \"d23e05af-c71b-4c8d-ac33-0a0aafbf3e32\" (UID: \"d23e05af-c71b-4c8d-ac33-0a0aafbf3e32\") " Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.837864 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d23e05af-c71b-4c8d-ac33-0a0aafbf3e32-kube-api-access-bvdcf" (OuterVolumeSpecName: "kube-api-access-bvdcf") pod "d23e05af-c71b-4c8d-ac33-0a0aafbf3e32" (UID: "d23e05af-c71b-4c8d-ac33-0a0aafbf3e32"). InnerVolumeSpecName "kube-api-access-bvdcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:23 crc kubenswrapper[4969]: I1004 08:33:23.934475 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvdcf\" (UniqueName: \"kubernetes.io/projected/d23e05af-c71b-4c8d-ac33-0a0aafbf3e32-kube-api-access-bvdcf\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.334471 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"01226baf93ec7900f3fa3b38a0cfca3644c4e5751e18c361d0b3d802d9ba2ac8"} Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.334734 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"8611389cf806eb2660f989b83af621cd086b1d6c2dcbc49f15da15c165266749"} Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.334744 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"be78e445bca170b20d1a51ab58a00fe9cc3fe1058ea86bdd16b5a046e1e402d4"} Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.338633 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-5zcjl" Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.338637 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-5zcjl" event={"ID":"d23e05af-c71b-4c8d-ac33-0a0aafbf3e32","Type":"ContainerDied","Data":"20b8c0fe6f807eb1824f5d72965832f3316b846b286e699ae7bc3d4c1ec05d60"} Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.338704 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20b8c0fe6f807eb1824f5d72965832f3316b846b286e699ae7bc3d4c1ec05d60" Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.701342 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-5c31-account-create-cqhsm" Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.747332 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbvwt\" (UniqueName: \"kubernetes.io/projected/472bf752-8d20-4b60-82f8-71c73e4f8843-kube-api-access-dbvwt\") pod \"472bf752-8d20-4b60-82f8-71c73e4f8843\" (UID: \"472bf752-8d20-4b60-82f8-71c73e4f8843\") " Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.753643 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/472bf752-8d20-4b60-82f8-71c73e4f8843-kube-api-access-dbvwt" (OuterVolumeSpecName: "kube-api-access-dbvwt") pod "472bf752-8d20-4b60-82f8-71c73e4f8843" (UID: "472bf752-8d20-4b60-82f8-71c73e4f8843"). InnerVolumeSpecName "kube-api-access-dbvwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:24 crc kubenswrapper[4969]: I1004 08:33:24.849033 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbvwt\" (UniqueName: \"kubernetes.io/projected/472bf752-8d20-4b60-82f8-71c73e4f8843-kube-api-access-dbvwt\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.202480 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fms7j" podUID="452b4c0a-fcb1-48ed-8c3a-87c34904a0a5" containerName="ovn-controller" probeResult="failure" output=< Oct 04 08:33:25 crc kubenswrapper[4969]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 08:33:25 crc kubenswrapper[4969]: > Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.353500 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"08c5c044c0edd64814c52aae354251449f3e5623969ee4798219fcc97a712f1f"} Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.353542 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"8da438d99e539a4ecb399be810b49abc4c9d6b6e4c5564fab49141a529549248"} Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.353551 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"90081444e8d3b1f3540fe64f355f51c4e396941bfdf5826c8a4143884ac896e3"} Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.353561 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"37a34ee4f2825acbe031308f26de2ebe5411487f6585544d657790af6a41db48"} Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.353568 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"6f212609b3539d9b23764fc39a9c60b11ef39b74ee72cdeb99b3fc7b8d17e02e"} Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.355491 4969 generic.go:334] "Generic (PLEG): container finished" podID="6eaebec4-b71b-409d-b91e-1993605ce209" containerID="8b1f66b1e50805e2ec1fadbeadbbdb658b9a97373b0469c6ef10e190672e5c1c" exitCode=0 Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.355581 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6eaebec4-b71b-409d-b91e-1993605ce209","Type":"ContainerDied","Data":"8b1f66b1e50805e2ec1fadbeadbbdb658b9a97373b0469c6ef10e190672e5c1c"} Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.358307 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-5c31-account-create-cqhsm" event={"ID":"472bf752-8d20-4b60-82f8-71c73e4f8843","Type":"ContainerDied","Data":"f88489c00e2c6768c7078489a1b70eb08300074207a6647cbf242ce2a82db1b5"} Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.358336 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f88489c00e2c6768c7078489a1b70eb08300074207a6647cbf242ce2a82db1b5" Oct 04 08:33:25 crc kubenswrapper[4969]: I1004 08:33:25.358665 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-5c31-account-create-cqhsm" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.371325 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6eaebec4-b71b-409d-b91e-1993605ce209","Type":"ContainerStarted","Data":"eadc989bb8e449b5c0280b3c9f133dc07201cbfffbd1c533c6a2d8bae55469e2"} Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.372022 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.381923 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4","Type":"ContainerStarted","Data":"c2b243ce97791011a75e98640451da149d442593cf0549d276ee325c7f537f34"} Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.422398 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=53.909544644 podStartE2EDuration="1m2.422363953s" podCreationTimestamp="2025-10-04 08:32:24 +0000 UTC" firstStartedPulling="2025-10-04 08:32:40.797337868 +0000 UTC m=+988.551606682" lastFinishedPulling="2025-10-04 08:32:49.310157187 +0000 UTC m=+997.064425991" observedRunningTime="2025-10-04 08:33:26.411286997 +0000 UTC m=+1034.165555851" watchObservedRunningTime="2025-10-04 08:33:26.422363953 +0000 UTC m=+1034.176632807" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.467823 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.750806111 podStartE2EDuration="24.467789313s" podCreationTimestamp="2025-10-04 08:33:02 +0000 UTC" firstStartedPulling="2025-10-04 08:33:20.400081042 +0000 UTC m=+1028.154349866" lastFinishedPulling="2025-10-04 08:33:24.117064254 +0000 UTC m=+1031.871333068" observedRunningTime="2025-10-04 08:33:26.456605415 +0000 UTC m=+1034.210874269" watchObservedRunningTime="2025-10-04 08:33:26.467789313 +0000 UTC m=+1034.222058137" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.734485 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-0476-account-create-vjvkm"] Oct 04 08:33:26 crc kubenswrapper[4969]: E1004 08:33:26.734992 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d6b82d-a2fa-463e-8075-04bf086f145b" containerName="mariadb-database-create" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.735021 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d6b82d-a2fa-463e-8075-04bf086f145b" containerName="mariadb-database-create" Oct 04 08:33:26 crc kubenswrapper[4969]: E1004 08:33:26.735067 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="472bf752-8d20-4b60-82f8-71c73e4f8843" containerName="mariadb-account-create" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.735076 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="472bf752-8d20-4b60-82f8-71c73e4f8843" containerName="mariadb-account-create" Oct 04 08:33:26 crc kubenswrapper[4969]: E1004 08:33:26.735100 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d23e05af-c71b-4c8d-ac33-0a0aafbf3e32" containerName="mariadb-database-create" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.735109 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="d23e05af-c71b-4c8d-ac33-0a0aafbf3e32" containerName="mariadb-database-create" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.735304 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="472bf752-8d20-4b60-82f8-71c73e4f8843" containerName="mariadb-account-create" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.735330 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d6b82d-a2fa-463e-8075-04bf086f145b" containerName="mariadb-database-create" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.735350 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="d23e05af-c71b-4c8d-ac33-0a0aafbf3e32" containerName="mariadb-database-create" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.736028 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0476-account-create-vjvkm" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.737770 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.745048 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cdb94bc5-xjwz9"] Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.750243 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.752985 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.753931 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0476-account-create-vjvkm"] Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.773557 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cdb94bc5-xjwz9"] Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.782471 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-swift-storage-0\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.782541 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-sb\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.782585 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zkn9\" (UniqueName: \"kubernetes.io/projected/63362185-ca5e-442d-8684-ee9a819ddb9e-kube-api-access-4zkn9\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.782610 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-nb\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.782670 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-svc\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.782698 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78k4s\" (UniqueName: \"kubernetes.io/projected/8f213477-11ae-42a6-b4e7-376453c4f64a-kube-api-access-78k4s\") pod \"glance-0476-account-create-vjvkm\" (UID: \"8f213477-11ae-42a6-b4e7-376453c4f64a\") " pod="openstack/glance-0476-account-create-vjvkm" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.782776 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-config\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.884725 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-svc\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.884805 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78k4s\" (UniqueName: \"kubernetes.io/projected/8f213477-11ae-42a6-b4e7-376453c4f64a-kube-api-access-78k4s\") pod \"glance-0476-account-create-vjvkm\" (UID: \"8f213477-11ae-42a6-b4e7-376453c4f64a\") " pod="openstack/glance-0476-account-create-vjvkm" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.884898 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-config\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.884930 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-swift-storage-0\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.884965 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-sb\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.885002 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zkn9\" (UniqueName: \"kubernetes.io/projected/63362185-ca5e-442d-8684-ee9a819ddb9e-kube-api-access-4zkn9\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.885027 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-nb\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.885630 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-svc\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.886171 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-sb\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.886171 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-config\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.886279 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-nb\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.886816 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-swift-storage-0\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.902851 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zkn9\" (UniqueName: \"kubernetes.io/projected/63362185-ca5e-442d-8684-ee9a819ddb9e-kube-api-access-4zkn9\") pod \"dnsmasq-dns-6cdb94bc5-xjwz9\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:26 crc kubenswrapper[4969]: I1004 08:33:26.904217 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78k4s\" (UniqueName: \"kubernetes.io/projected/8f213477-11ae-42a6-b4e7-376453c4f64a-kube-api-access-78k4s\") pod \"glance-0476-account-create-vjvkm\" (UID: \"8f213477-11ae-42a6-b4e7-376453c4f64a\") " pod="openstack/glance-0476-account-create-vjvkm" Oct 04 08:33:27 crc kubenswrapper[4969]: I1004 08:33:27.052513 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0476-account-create-vjvkm" Oct 04 08:33:27 crc kubenswrapper[4969]: I1004 08:33:27.069084 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:27 crc kubenswrapper[4969]: W1004 08:33:27.579537 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63362185_ca5e_442d_8684_ee9a819ddb9e.slice/crio-e290e1a74c6f81c717ff9e40b836840f302b294a6fde765d1f611347df0076b0 WatchSource:0}: Error finding container e290e1a74c6f81c717ff9e40b836840f302b294a6fde765d1f611347df0076b0: Status 404 returned error can't find the container with id e290e1a74c6f81c717ff9e40b836840f302b294a6fde765d1f611347df0076b0 Oct 04 08:33:27 crc kubenswrapper[4969]: I1004 08:33:27.586738 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0476-account-create-vjvkm"] Oct 04 08:33:27 crc kubenswrapper[4969]: I1004 08:33:27.611142 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cdb94bc5-xjwz9"] Oct 04 08:33:28 crc kubenswrapper[4969]: I1004 08:33:28.411828 4969 generic.go:334] "Generic (PLEG): container finished" podID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerID="aa7b8a719157409a83ce94cd7a2e1bd6a4d71b6187f5c255388dd918f6447a3f" exitCode=0 Oct 04 08:33:28 crc kubenswrapper[4969]: I1004 08:33:28.411951 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" event={"ID":"63362185-ca5e-442d-8684-ee9a819ddb9e","Type":"ContainerDied","Data":"aa7b8a719157409a83ce94cd7a2e1bd6a4d71b6187f5c255388dd918f6447a3f"} Oct 04 08:33:28 crc kubenswrapper[4969]: I1004 08:33:28.412316 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" event={"ID":"63362185-ca5e-442d-8684-ee9a819ddb9e","Type":"ContainerStarted","Data":"e290e1a74c6f81c717ff9e40b836840f302b294a6fde765d1f611347df0076b0"} Oct 04 08:33:28 crc kubenswrapper[4969]: I1004 08:33:28.415524 4969 generic.go:334] "Generic (PLEG): container finished" podID="8f213477-11ae-42a6-b4e7-376453c4f64a" containerID="8558501ad2698336951a93011571b0c299db2d03fb0cad4750355cfce6089045" exitCode=0 Oct 04 08:33:28 crc kubenswrapper[4969]: I1004 08:33:28.415590 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0476-account-create-vjvkm" event={"ID":"8f213477-11ae-42a6-b4e7-376453c4f64a","Type":"ContainerDied","Data":"8558501ad2698336951a93011571b0c299db2d03fb0cad4750355cfce6089045"} Oct 04 08:33:28 crc kubenswrapper[4969]: I1004 08:33:28.415630 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0476-account-create-vjvkm" event={"ID":"8f213477-11ae-42a6-b4e7-376453c4f64a","Type":"ContainerStarted","Data":"9626056ac6b370d819d5fb6cd57ccc681fcd007b059f224adf3ff64309b000f2"} Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.430578 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" event={"ID":"63362185-ca5e-442d-8684-ee9a819ddb9e","Type":"ContainerStarted","Data":"58defe36623f82a68e5be64593dce1537b2a58a430849b278909c50ee12e9745"} Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.433250 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.435893 4969 generic.go:334] "Generic (PLEG): container finished" podID="1547b86a-ce65-4135-a8a5-957017c2271c" containerID="7293ab8d08343f9c68c67fb2c9a6ca5eae2fbf92ee2915c9b9c3d3620b5a04f8" exitCode=0 Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.435986 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1547b86a-ce65-4135-a8a5-957017c2271c","Type":"ContainerDied","Data":"7293ab8d08343f9c68c67fb2c9a6ca5eae2fbf92ee2915c9b9c3d3620b5a04f8"} Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.490722 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" podStartSLOduration=3.4906987259999998 podStartE2EDuration="3.490698726s" podCreationTimestamp="2025-10-04 08:33:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:33:29.459231263 +0000 UTC m=+1037.213500117" watchObservedRunningTime="2025-10-04 08:33:29.490698726 +0000 UTC m=+1037.244967550" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.727363 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-34b9-account-create-z7dgt"] Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.728434 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-34b9-account-create-z7dgt" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.730265 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.743544 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-34b9-account-create-z7dgt"] Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.814850 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0476-account-create-vjvkm" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.849061 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9rtt\" (UniqueName: \"kubernetes.io/projected/8187bdb9-d868-4c45-bb7e-df0c38f70e9b-kube-api-access-w9rtt\") pod \"keystone-34b9-account-create-z7dgt\" (UID: \"8187bdb9-d868-4c45-bb7e-df0c38f70e9b\") " pod="openstack/keystone-34b9-account-create-z7dgt" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.916839 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a8d7-account-create-xgxvr"] Oct 04 08:33:29 crc kubenswrapper[4969]: E1004 08:33:29.917189 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f213477-11ae-42a6-b4e7-376453c4f64a" containerName="mariadb-account-create" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.917210 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f213477-11ae-42a6-b4e7-376453c4f64a" containerName="mariadb-account-create" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.917408 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f213477-11ae-42a6-b4e7-376453c4f64a" containerName="mariadb-account-create" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.918055 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a8d7-account-create-xgxvr" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.922404 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.925658 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a8d7-account-create-xgxvr"] Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.950357 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78k4s\" (UniqueName: \"kubernetes.io/projected/8f213477-11ae-42a6-b4e7-376453c4f64a-kube-api-access-78k4s\") pod \"8f213477-11ae-42a6-b4e7-376453c4f64a\" (UID: \"8f213477-11ae-42a6-b4e7-376453c4f64a\") " Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.950876 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9rtt\" (UniqueName: \"kubernetes.io/projected/8187bdb9-d868-4c45-bb7e-df0c38f70e9b-kube-api-access-w9rtt\") pod \"keystone-34b9-account-create-z7dgt\" (UID: \"8187bdb9-d868-4c45-bb7e-df0c38f70e9b\") " pod="openstack/keystone-34b9-account-create-z7dgt" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.955038 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f213477-11ae-42a6-b4e7-376453c4f64a-kube-api-access-78k4s" (OuterVolumeSpecName: "kube-api-access-78k4s") pod "8f213477-11ae-42a6-b4e7-376453c4f64a" (UID: "8f213477-11ae-42a6-b4e7-376453c4f64a"). InnerVolumeSpecName "kube-api-access-78k4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:29 crc kubenswrapper[4969]: I1004 08:33:29.968095 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9rtt\" (UniqueName: \"kubernetes.io/projected/8187bdb9-d868-4c45-bb7e-df0c38f70e9b-kube-api-access-w9rtt\") pod \"keystone-34b9-account-create-z7dgt\" (UID: \"8187bdb9-d868-4c45-bb7e-df0c38f70e9b\") " pod="openstack/keystone-34b9-account-create-z7dgt" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.053306 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4m4r\" (UniqueName: \"kubernetes.io/projected/3293da46-00e5-4183-a2ed-7c702651f3cc-kube-api-access-r4m4r\") pod \"placement-a8d7-account-create-xgxvr\" (UID: \"3293da46-00e5-4183-a2ed-7c702651f3cc\") " pod="openstack/placement-a8d7-account-create-xgxvr" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.053626 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78k4s\" (UniqueName: \"kubernetes.io/projected/8f213477-11ae-42a6-b4e7-376453c4f64a-kube-api-access-78k4s\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.112705 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-34b9-account-create-z7dgt" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.154635 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4m4r\" (UniqueName: \"kubernetes.io/projected/3293da46-00e5-4183-a2ed-7c702651f3cc-kube-api-access-r4m4r\") pod \"placement-a8d7-account-create-xgxvr\" (UID: \"3293da46-00e5-4183-a2ed-7c702651f3cc\") " pod="openstack/placement-a8d7-account-create-xgxvr" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.174303 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4m4r\" (UniqueName: \"kubernetes.io/projected/3293da46-00e5-4183-a2ed-7c702651f3cc-kube-api-access-r4m4r\") pod \"placement-a8d7-account-create-xgxvr\" (UID: \"3293da46-00e5-4183-a2ed-7c702651f3cc\") " pod="openstack/placement-a8d7-account-create-xgxvr" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.195358 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fms7j" podUID="452b4c0a-fcb1-48ed-8c3a-87c34904a0a5" containerName="ovn-controller" probeResult="failure" output=< Oct 04 08:33:30 crc kubenswrapper[4969]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 08:33:30 crc kubenswrapper[4969]: > Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.212943 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.244290 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a8d7-account-create-xgxvr" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.249698 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-p9gfr" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.481092 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fms7j-config-vgfh8"] Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.484876 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.486610 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.488683 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fms7j-config-vgfh8"] Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.492192 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1547b86a-ce65-4135-a8a5-957017c2271c","Type":"ContainerStarted","Data":"001e2338fea808e3665ffd16d9def28352878cb322c3ca5065371fdfd6537ced"} Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.492512 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.494150 4969 generic.go:334] "Generic (PLEG): container finished" podID="28264a08-8168-419d-a20d-8f94a5c4ed77" containerID="6cfa4591c67cb36af29eb68362609a083c92363f9ed19aba47fa884f35f7da4c" exitCode=0 Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.494219 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"28264a08-8168-419d-a20d-8f94a5c4ed77","Type":"ContainerDied","Data":"6cfa4591c67cb36af29eb68362609a083c92363f9ed19aba47fa884f35f7da4c"} Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.498805 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0476-account-create-vjvkm" event={"ID":"8f213477-11ae-42a6-b4e7-376453c4f64a","Type":"ContainerDied","Data":"9626056ac6b370d819d5fb6cd57ccc681fcd007b059f224adf3ff64309b000f2"} Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.498847 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9626056ac6b370d819d5fb6cd57ccc681fcd007b059f224adf3ff64309b000f2" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.500006 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0476-account-create-vjvkm" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.583178 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=57.659788862 podStartE2EDuration="1m6.583159831s" podCreationTimestamp="2025-10-04 08:32:24 +0000 UTC" firstStartedPulling="2025-10-04 08:32:40.794571579 +0000 UTC m=+988.548840393" lastFinishedPulling="2025-10-04 08:32:49.717942538 +0000 UTC m=+997.472211362" observedRunningTime="2025-10-04 08:33:30.563500662 +0000 UTC m=+1038.317769496" watchObservedRunningTime="2025-10-04 08:33:30.583159831 +0000 UTC m=+1038.337428655" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.592972 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-34b9-account-create-z7dgt"] Oct 04 08:33:30 crc kubenswrapper[4969]: W1004 08:33:30.598172 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8187bdb9_d868_4c45_bb7e_df0c38f70e9b.slice/crio-5509975e6fbd79ea8bfbb03de2410e7ddf3cd29de86afb626d8e3d7a3ba5c47c WatchSource:0}: Error finding container 5509975e6fbd79ea8bfbb03de2410e7ddf3cd29de86afb626d8e3d7a3ba5c47c: Status 404 returned error can't find the container with id 5509975e6fbd79ea8bfbb03de2410e7ddf3cd29de86afb626d8e3d7a3ba5c47c Oct 04 08:33:30 crc kubenswrapper[4969]: W1004 08:33:30.675661 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3293da46_00e5_4183_a2ed_7c702651f3cc.slice/crio-ee5ab08d1ff00a8dc204ed8d76184dc0ca3cb350f16ae89cab597d0364cbd394 WatchSource:0}: Error finding container ee5ab08d1ff00a8dc204ed8d76184dc0ca3cb350f16ae89cab597d0364cbd394: Status 404 returned error can't find the container with id ee5ab08d1ff00a8dc204ed8d76184dc0ca3cb350f16ae89cab597d0364cbd394 Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.678808 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.678849 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmh4r\" (UniqueName: \"kubernetes.io/projected/c95a5b47-c935-46e5-9dcc-d3352aac8b98-kube-api-access-hmh4r\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.678908 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run-ovn\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.679047 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-log-ovn\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.679052 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a8d7-account-create-xgxvr"] Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.679117 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-additional-scripts\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.679161 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-scripts\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.781135 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-scripts\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.781192 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.781222 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmh4r\" (UniqueName: \"kubernetes.io/projected/c95a5b47-c935-46e5-9dcc-d3352aac8b98-kube-api-access-hmh4r\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.781267 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run-ovn\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.781356 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-log-ovn\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.781390 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-additional-scripts\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.781524 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run-ovn\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.781589 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-log-ovn\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.781623 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.783026 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-scripts\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.784836 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-additional-scripts\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.801521 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmh4r\" (UniqueName: \"kubernetes.io/projected/c95a5b47-c935-46e5-9dcc-d3352aac8b98-kube-api-access-hmh4r\") pod \"ovn-controller-fms7j-config-vgfh8\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:30 crc kubenswrapper[4969]: I1004 08:33:30.958929 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.418521 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fms7j-config-vgfh8"] Oct 04 08:33:31 crc kubenswrapper[4969]: W1004 08:33:31.427494 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc95a5b47_c935_46e5_9dcc_d3352aac8b98.slice/crio-50ac3b5f0bf29279f0b3d705a2ea945881b8d06a3961108a637bff25f4abbf6c WatchSource:0}: Error finding container 50ac3b5f0bf29279f0b3d705a2ea945881b8d06a3961108a637bff25f4abbf6c: Status 404 returned error can't find the container with id 50ac3b5f0bf29279f0b3d705a2ea945881b8d06a3961108a637bff25f4abbf6c Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.509781 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"28264a08-8168-419d-a20d-8f94a5c4ed77","Type":"ContainerStarted","Data":"e824dbb3f6c6ed2205d6e595f08051bc9432f5ca83e9dea6b7212036eab208b2"} Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.510652 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.521506 4969 generic.go:334] "Generic (PLEG): container finished" podID="3293da46-00e5-4183-a2ed-7c702651f3cc" containerID="72d3eea0e6a149ea0314050fc3f297c8f1a2f61001947c5edd930688eaf47004" exitCode=0 Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.521566 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a8d7-account-create-xgxvr" event={"ID":"3293da46-00e5-4183-a2ed-7c702651f3cc","Type":"ContainerDied","Data":"72d3eea0e6a149ea0314050fc3f297c8f1a2f61001947c5edd930688eaf47004"} Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.521587 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a8d7-account-create-xgxvr" event={"ID":"3293da46-00e5-4183-a2ed-7c702651f3cc","Type":"ContainerStarted","Data":"ee5ab08d1ff00a8dc204ed8d76184dc0ca3cb350f16ae89cab597d0364cbd394"} Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.526818 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fms7j-config-vgfh8" event={"ID":"c95a5b47-c935-46e5-9dcc-d3352aac8b98","Type":"ContainerStarted","Data":"50ac3b5f0bf29279f0b3d705a2ea945881b8d06a3961108a637bff25f4abbf6c"} Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.532983 4969 generic.go:334] "Generic (PLEG): container finished" podID="8187bdb9-d868-4c45-bb7e-df0c38f70e9b" containerID="9644a9f250a026f90c923338b9d3be04047805319fc90514b03999744717a5e5" exitCode=0 Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.533023 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-34b9-account-create-z7dgt" event={"ID":"8187bdb9-d868-4c45-bb7e-df0c38f70e9b","Type":"ContainerDied","Data":"9644a9f250a026f90c923338b9d3be04047805319fc90514b03999744717a5e5"} Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.533969 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-34b9-account-create-z7dgt" event={"ID":"8187bdb9-d868-4c45-bb7e-df0c38f70e9b","Type":"ContainerStarted","Data":"5509975e6fbd79ea8bfbb03de2410e7ddf3cd29de86afb626d8e3d7a3ba5c47c"} Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.560416 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-notifications-server-0" podStartSLOduration=57.486627527 podStartE2EDuration="1m6.560395759s" podCreationTimestamp="2025-10-04 08:32:25 +0000 UTC" firstStartedPulling="2025-10-04 08:32:41.088930747 +0000 UTC m=+988.843199561" lastFinishedPulling="2025-10-04 08:32:50.162698979 +0000 UTC m=+997.916967793" observedRunningTime="2025-10-04 08:33:31.546981246 +0000 UTC m=+1039.301250070" watchObservedRunningTime="2025-10-04 08:33:31.560395759 +0000 UTC m=+1039.314664573" Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.872055 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-2t48v"] Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.876279 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.879200 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.879339 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vmpm7" Oct 04 08:33:31 crc kubenswrapper[4969]: I1004 08:33:31.887313 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-2t48v"] Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.002897 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-config-data\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.002947 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-db-sync-config-data\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.003010 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh7fp\" (UniqueName: \"kubernetes.io/projected/ed6a88af-0fc2-45f8-b584-451c87b14ae6-kube-api-access-dh7fp\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.003076 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-combined-ca-bundle\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.105529 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-config-data\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.105584 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-db-sync-config-data\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.105650 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh7fp\" (UniqueName: \"kubernetes.io/projected/ed6a88af-0fc2-45f8-b584-451c87b14ae6-kube-api-access-dh7fp\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.105712 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-combined-ca-bundle\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.111692 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-combined-ca-bundle\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.111781 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-db-sync-config-data\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.113015 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-config-data\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.124102 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh7fp\" (UniqueName: \"kubernetes.io/projected/ed6a88af-0fc2-45f8-b584-451c87b14ae6-kube-api-access-dh7fp\") pod \"glance-db-sync-2t48v\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.205555 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.546033 4969 generic.go:334] "Generic (PLEG): container finished" podID="c95a5b47-c935-46e5-9dcc-d3352aac8b98" containerID="85771b674ebd673fcaeda8fa43cea00728395310592d2f7c048d0767b1c51055" exitCode=0 Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.546180 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fms7j-config-vgfh8" event={"ID":"c95a5b47-c935-46e5-9dcc-d3352aac8b98","Type":"ContainerDied","Data":"85771b674ebd673fcaeda8fa43cea00728395310592d2f7c048d0767b1c51055"} Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.879336 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-2t48v"] Oct 04 08:33:32 crc kubenswrapper[4969]: I1004 08:33:32.991500 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a8d7-account-create-xgxvr" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.069034 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-34b9-account-create-z7dgt" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.133533 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4m4r\" (UniqueName: \"kubernetes.io/projected/3293da46-00e5-4183-a2ed-7c702651f3cc-kube-api-access-r4m4r\") pod \"3293da46-00e5-4183-a2ed-7c702651f3cc\" (UID: \"3293da46-00e5-4183-a2ed-7c702651f3cc\") " Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.139541 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3293da46-00e5-4183-a2ed-7c702651f3cc-kube-api-access-r4m4r" (OuterVolumeSpecName: "kube-api-access-r4m4r") pod "3293da46-00e5-4183-a2ed-7c702651f3cc" (UID: "3293da46-00e5-4183-a2ed-7c702651f3cc"). InnerVolumeSpecName "kube-api-access-r4m4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.190734 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.194213 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.235025 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rtt\" (UniqueName: \"kubernetes.io/projected/8187bdb9-d868-4c45-bb7e-df0c38f70e9b-kube-api-access-w9rtt\") pod \"8187bdb9-d868-4c45-bb7e-df0c38f70e9b\" (UID: \"8187bdb9-d868-4c45-bb7e-df0c38f70e9b\") " Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.235773 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4m4r\" (UniqueName: \"kubernetes.io/projected/3293da46-00e5-4183-a2ed-7c702651f3cc-kube-api-access-r4m4r\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.239368 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8187bdb9-d868-4c45-bb7e-df0c38f70e9b-kube-api-access-w9rtt" (OuterVolumeSpecName: "kube-api-access-w9rtt") pod "8187bdb9-d868-4c45-bb7e-df0c38f70e9b" (UID: "8187bdb9-d868-4c45-bb7e-df0c38f70e9b"). InnerVolumeSpecName "kube-api-access-w9rtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.337186 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rtt\" (UniqueName: \"kubernetes.io/projected/8187bdb9-d868-4c45-bb7e-df0c38f70e9b-kube-api-access-w9rtt\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.567907 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a8d7-account-create-xgxvr" event={"ID":"3293da46-00e5-4183-a2ed-7c702651f3cc","Type":"ContainerDied","Data":"ee5ab08d1ff00a8dc204ed8d76184dc0ca3cb350f16ae89cab597d0364cbd394"} Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.567965 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee5ab08d1ff00a8dc204ed8d76184dc0ca3cb350f16ae89cab597d0364cbd394" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.568103 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a8d7-account-create-xgxvr" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.573125 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-34b9-account-create-z7dgt" event={"ID":"8187bdb9-d868-4c45-bb7e-df0c38f70e9b","Type":"ContainerDied","Data":"5509975e6fbd79ea8bfbb03de2410e7ddf3cd29de86afb626d8e3d7a3ba5c47c"} Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.573212 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5509975e6fbd79ea8bfbb03de2410e7ddf3cd29de86afb626d8e3d7a3ba5c47c" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.573316 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-34b9-account-create-z7dgt" Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.580112 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2t48v" event={"ID":"ed6a88af-0fc2-45f8-b584-451c87b14ae6","Type":"ContainerStarted","Data":"b76536b25ffe2a368483a084c0d10400c4329b23be844cc2145104c92ab61ec2"} Oct 04 08:33:33 crc kubenswrapper[4969]: I1004 08:33:33.581293 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.021776 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.152616 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmh4r\" (UniqueName: \"kubernetes.io/projected/c95a5b47-c935-46e5-9dcc-d3352aac8b98-kube-api-access-hmh4r\") pod \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.152671 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-log-ovn\") pod \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.152792 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c95a5b47-c935-46e5-9dcc-d3352aac8b98" (UID: "c95a5b47-c935-46e5-9dcc-d3352aac8b98"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.152805 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-additional-scripts\") pod \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.152905 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run\") pod \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.152951 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run-ovn\") pod \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.153024 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run" (OuterVolumeSpecName: "var-run") pod "c95a5b47-c935-46e5-9dcc-d3352aac8b98" (UID: "c95a5b47-c935-46e5-9dcc-d3352aac8b98"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.153074 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c95a5b47-c935-46e5-9dcc-d3352aac8b98" (UID: "c95a5b47-c935-46e5-9dcc-d3352aac8b98"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.153177 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-scripts\") pod \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\" (UID: \"c95a5b47-c935-46e5-9dcc-d3352aac8b98\") " Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.153658 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c95a5b47-c935-46e5-9dcc-d3352aac8b98" (UID: "c95a5b47-c935-46e5-9dcc-d3352aac8b98"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.154017 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-scripts" (OuterVolumeSpecName: "scripts") pod "c95a5b47-c935-46e5-9dcc-d3352aac8b98" (UID: "c95a5b47-c935-46e5-9dcc-d3352aac8b98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.154949 4969 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.154977 4969 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.154988 4969 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.154998 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95a5b47-c935-46e5-9dcc-d3352aac8b98-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.155007 4969 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c95a5b47-c935-46e5-9dcc-d3352aac8b98-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.158006 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c95a5b47-c935-46e5-9dcc-d3352aac8b98-kube-api-access-hmh4r" (OuterVolumeSpecName: "kube-api-access-hmh4r") pod "c95a5b47-c935-46e5-9dcc-d3352aac8b98" (UID: "c95a5b47-c935-46e5-9dcc-d3352aac8b98"). InnerVolumeSpecName "kube-api-access-hmh4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.257501 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmh4r\" (UniqueName: \"kubernetes.io/projected/c95a5b47-c935-46e5-9dcc-d3352aac8b98-kube-api-access-hmh4r\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.590362 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fms7j-config-vgfh8" event={"ID":"c95a5b47-c935-46e5-9dcc-d3352aac8b98","Type":"ContainerDied","Data":"50ac3b5f0bf29279f0b3d705a2ea945881b8d06a3961108a637bff25f4abbf6c"} Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.590416 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50ac3b5f0bf29279f0b3d705a2ea945881b8d06a3961108a637bff25f4abbf6c" Oct 04 08:33:34 crc kubenswrapper[4969]: I1004 08:33:34.590450 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fms7j-config-vgfh8" Oct 04 08:33:35 crc kubenswrapper[4969]: I1004 08:33:35.127229 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fms7j-config-vgfh8"] Oct 04 08:33:35 crc kubenswrapper[4969]: I1004 08:33:35.132257 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fms7j-config-vgfh8"] Oct 04 08:33:35 crc kubenswrapper[4969]: I1004 08:33:35.191603 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-fms7j" Oct 04 08:33:36 crc kubenswrapper[4969]: I1004 08:33:36.202740 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="6eaebec4-b71b-409d-b91e-1993605ce209" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 04 08:33:36 crc kubenswrapper[4969]: I1004 08:33:36.705700 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 08:33:36 crc kubenswrapper[4969]: I1004 08:33:36.705966 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="prometheus" containerID="cri-o://0fc68e0c8ef687a06bb8efb6b04020362453bde79bf724ff5315525cd9b3fc7b" gracePeriod=600 Oct 04 08:33:36 crc kubenswrapper[4969]: I1004 08:33:36.706034 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="thanos-sidecar" containerID="cri-o://2edabe18cb5eb64471d46ba83a6f332ca980c9481e667db15cafd0f1131e0a64" gracePeriod=600 Oct 04 08:33:36 crc kubenswrapper[4969]: I1004 08:33:36.706107 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="config-reloader" containerID="cri-o://488038e962c9c490bfc090e32286e992e959221611857e4f0b3009eb80ffb8ac" gracePeriod=600 Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.069588 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c95a5b47-c935-46e5-9dcc-d3352aac8b98" path="/var/lib/kubelet/pods/c95a5b47-c935-46e5-9dcc-d3352aac8b98/volumes" Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.070945 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.123384 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79795ddc65-wj2fb"] Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.123623 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" podUID="1de7855b-90f1-46ae-a85c-9e292224b87e" containerName="dnsmasq-dns" containerID="cri-o://6149164f902be6d8b639a5f942ae0d60f89cff145bcca988a8150693ef37f0a0" gracePeriod=10 Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.318073 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" podUID="1de7855b-90f1-46ae-a85c-9e292224b87e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: connect: connection refused" Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.614934 4969 generic.go:334] "Generic (PLEG): container finished" podID="1de7855b-90f1-46ae-a85c-9e292224b87e" containerID="6149164f902be6d8b639a5f942ae0d60f89cff145bcca988a8150693ef37f0a0" exitCode=0 Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.615031 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" event={"ID":"1de7855b-90f1-46ae-a85c-9e292224b87e","Type":"ContainerDied","Data":"6149164f902be6d8b639a5f942ae0d60f89cff145bcca988a8150693ef37f0a0"} Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.624735 4969 generic.go:334] "Generic (PLEG): container finished" podID="49069f68-580b-41f3-9504-6e8b722459f7" containerID="2edabe18cb5eb64471d46ba83a6f332ca980c9481e667db15cafd0f1131e0a64" exitCode=0 Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.624773 4969 generic.go:334] "Generic (PLEG): container finished" podID="49069f68-580b-41f3-9504-6e8b722459f7" containerID="488038e962c9c490bfc090e32286e992e959221611857e4f0b3009eb80ffb8ac" exitCode=0 Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.624785 4969 generic.go:334] "Generic (PLEG): container finished" podID="49069f68-580b-41f3-9504-6e8b722459f7" containerID="0fc68e0c8ef687a06bb8efb6b04020362453bde79bf724ff5315525cd9b3fc7b" exitCode=0 Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.624805 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerDied","Data":"2edabe18cb5eb64471d46ba83a6f332ca980c9481e667db15cafd0f1131e0a64"} Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.624834 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerDied","Data":"488038e962c9c490bfc090e32286e992e959221611857e4f0b3009eb80ffb8ac"} Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.624846 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerDied","Data":"0fc68e0c8ef687a06bb8efb6b04020362453bde79bf724ff5315525cd9b3fc7b"} Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.902845 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:37 crc kubenswrapper[4969]: I1004 08:33:37.912136 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.030099 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-web-config\") pod \"49069f68-580b-41f3-9504-6e8b722459f7\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.030361 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-thanos-prometheus-http-client-file\") pod \"49069f68-580b-41f3-9504-6e8b722459f7\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.030516 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"49069f68-580b-41f3-9504-6e8b722459f7\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.030552 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbcjh\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-kube-api-access-vbcjh\") pod \"49069f68-580b-41f3-9504-6e8b722459f7\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.030579 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/49069f68-580b-41f3-9504-6e8b722459f7-config-out\") pod \"49069f68-580b-41f3-9504-6e8b722459f7\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.030604 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/49069f68-580b-41f3-9504-6e8b722459f7-prometheus-metric-storage-rulefiles-0\") pod \"49069f68-580b-41f3-9504-6e8b722459f7\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.030679 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-dns-svc\") pod \"1de7855b-90f1-46ae-a85c-9e292224b87e\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.030713 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-nb\") pod \"1de7855b-90f1-46ae-a85c-9e292224b87e\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.031548 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-tls-assets\") pod \"49069f68-580b-41f3-9504-6e8b722459f7\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.031603 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kd6r8\" (UniqueName: \"kubernetes.io/projected/1de7855b-90f1-46ae-a85c-9e292224b87e-kube-api-access-kd6r8\") pod \"1de7855b-90f1-46ae-a85c-9e292224b87e\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.031665 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-config\") pod \"49069f68-580b-41f3-9504-6e8b722459f7\" (UID: \"49069f68-580b-41f3-9504-6e8b722459f7\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.031724 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-config\") pod \"1de7855b-90f1-46ae-a85c-9e292224b87e\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.031752 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-sb\") pod \"1de7855b-90f1-46ae-a85c-9e292224b87e\" (UID: \"1de7855b-90f1-46ae-a85c-9e292224b87e\") " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.031150 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49069f68-580b-41f3-9504-6e8b722459f7-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "49069f68-580b-41f3-9504-6e8b722459f7" (UID: "49069f68-580b-41f3-9504-6e8b722459f7"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.032323 4969 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/49069f68-580b-41f3-9504-6e8b722459f7-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.036566 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49069f68-580b-41f3-9504-6e8b722459f7-config-out" (OuterVolumeSpecName: "config-out") pod "49069f68-580b-41f3-9504-6e8b722459f7" (UID: "49069f68-580b-41f3-9504-6e8b722459f7"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.036573 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-kube-api-access-vbcjh" (OuterVolumeSpecName: "kube-api-access-vbcjh") pod "49069f68-580b-41f3-9504-6e8b722459f7" (UID: "49069f68-580b-41f3-9504-6e8b722459f7"). InnerVolumeSpecName "kube-api-access-vbcjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.037995 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "49069f68-580b-41f3-9504-6e8b722459f7" (UID: "49069f68-580b-41f3-9504-6e8b722459f7"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.038049 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-config" (OuterVolumeSpecName: "config") pod "49069f68-580b-41f3-9504-6e8b722459f7" (UID: "49069f68-580b-41f3-9504-6e8b722459f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.038584 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de7855b-90f1-46ae-a85c-9e292224b87e-kube-api-access-kd6r8" (OuterVolumeSpecName: "kube-api-access-kd6r8") pod "1de7855b-90f1-46ae-a85c-9e292224b87e" (UID: "1de7855b-90f1-46ae-a85c-9e292224b87e"). InnerVolumeSpecName "kube-api-access-kd6r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.045557 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "49069f68-580b-41f3-9504-6e8b722459f7" (UID: "49069f68-580b-41f3-9504-6e8b722459f7"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.082809 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-web-config" (OuterVolumeSpecName: "web-config") pod "49069f68-580b-41f3-9504-6e8b722459f7" (UID: "49069f68-580b-41f3-9504-6e8b722459f7"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.087132 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "49069f68-580b-41f3-9504-6e8b722459f7" (UID: "49069f68-580b-41f3-9504-6e8b722459f7"). InnerVolumeSpecName "pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.117154 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1de7855b-90f1-46ae-a85c-9e292224b87e" (UID: "1de7855b-90f1-46ae-a85c-9e292224b87e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.126204 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1de7855b-90f1-46ae-a85c-9e292224b87e" (UID: "1de7855b-90f1-46ae-a85c-9e292224b87e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.132083 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1de7855b-90f1-46ae-a85c-9e292224b87e" (UID: "1de7855b-90f1-46ae-a85c-9e292224b87e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.134455 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.134655 4969 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-web-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.134766 4969 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.134868 4969 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") on node \"crc\" " Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.135049 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbcjh\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-kube-api-access-vbcjh\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.135129 4969 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/49069f68-580b-41f3-9504-6e8b722459f7-config-out\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.135255 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.135334 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.135415 4969 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/49069f68-580b-41f3-9504-6e8b722459f7-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.135530 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kd6r8\" (UniqueName: \"kubernetes.io/projected/1de7855b-90f1-46ae-a85c-9e292224b87e-kube-api-access-kd6r8\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.135609 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/49069f68-580b-41f3-9504-6e8b722459f7-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.154770 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-config" (OuterVolumeSpecName: "config") pod "1de7855b-90f1-46ae-a85c-9e292224b87e" (UID: "1de7855b-90f1-46ae-a85c-9e292224b87e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.173800 4969 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.173956 4969 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a") on node "crc" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.237907 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de7855b-90f1-46ae-a85c-9e292224b87e-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.237943 4969 reconciler_common.go:293] "Volume detached for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.636232 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"49069f68-580b-41f3-9504-6e8b722459f7","Type":"ContainerDied","Data":"50089e4ac1d8024a9f1311d4600655ff4f6e53fade4e12109bc2d04d385edc65"} Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.636291 4969 scope.go:117] "RemoveContainer" containerID="2edabe18cb5eb64471d46ba83a6f332ca980c9481e667db15cafd0f1131e0a64" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.636467 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.646198 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" event={"ID":"1de7855b-90f1-46ae-a85c-9e292224b87e","Type":"ContainerDied","Data":"b5b8823e17beba8088dbdd8a7fb6ea4943737a8a988751f49dc5bb5862480c84"} Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.646255 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79795ddc65-wj2fb" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.683912 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.685641 4969 scope.go:117] "RemoveContainer" containerID="488038e962c9c490bfc090e32286e992e959221611857e4f0b3009eb80ffb8ac" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.696155 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.713005 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79795ddc65-wj2fb"] Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.721156 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79795ddc65-wj2fb"] Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727250 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 08:33:38 crc kubenswrapper[4969]: E1004 08:33:38.727606 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3293da46-00e5-4183-a2ed-7c702651f3cc" containerName="mariadb-account-create" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727624 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="3293da46-00e5-4183-a2ed-7c702651f3cc" containerName="mariadb-account-create" Oct 04 08:33:38 crc kubenswrapper[4969]: E1004 08:33:38.727635 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de7855b-90f1-46ae-a85c-9e292224b87e" containerName="init" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727641 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de7855b-90f1-46ae-a85c-9e292224b87e" containerName="init" Oct 04 08:33:38 crc kubenswrapper[4969]: E1004 08:33:38.727655 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c95a5b47-c935-46e5-9dcc-d3352aac8b98" containerName="ovn-config" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727660 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c95a5b47-c935-46e5-9dcc-d3352aac8b98" containerName="ovn-config" Oct 04 08:33:38 crc kubenswrapper[4969]: E1004 08:33:38.727684 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="config-reloader" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727690 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="config-reloader" Oct 04 08:33:38 crc kubenswrapper[4969]: E1004 08:33:38.727698 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8187bdb9-d868-4c45-bb7e-df0c38f70e9b" containerName="mariadb-account-create" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727703 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="8187bdb9-d868-4c45-bb7e-df0c38f70e9b" containerName="mariadb-account-create" Oct 04 08:33:38 crc kubenswrapper[4969]: E1004 08:33:38.727714 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de7855b-90f1-46ae-a85c-9e292224b87e" containerName="dnsmasq-dns" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727720 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de7855b-90f1-46ae-a85c-9e292224b87e" containerName="dnsmasq-dns" Oct 04 08:33:38 crc kubenswrapper[4969]: E1004 08:33:38.727728 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="prometheus" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727734 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="prometheus" Oct 04 08:33:38 crc kubenswrapper[4969]: E1004 08:33:38.727755 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="thanos-sidecar" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727760 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="thanos-sidecar" Oct 04 08:33:38 crc kubenswrapper[4969]: E1004 08:33:38.727767 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="init-config-reloader" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727774 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="init-config-reloader" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727907 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c95a5b47-c935-46e5-9dcc-d3352aac8b98" containerName="ovn-config" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727917 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="thanos-sidecar" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727932 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="3293da46-00e5-4183-a2ed-7c702651f3cc" containerName="mariadb-account-create" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727941 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="prometheus" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727955 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de7855b-90f1-46ae-a85c-9e292224b87e" containerName="dnsmasq-dns" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727964 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="8187bdb9-d868-4c45-bb7e-df0c38f70e9b" containerName="mariadb-account-create" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.727973 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="49069f68-580b-41f3-9504-6e8b722459f7" containerName="config-reloader" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.729125 4969 scope.go:117] "RemoveContainer" containerID="0fc68e0c8ef687a06bb8efb6b04020362453bde79bf724ff5315525cd9b3fc7b" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.729396 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.739580 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.739920 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.740191 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.740311 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-n8xt6" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.740566 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.740723 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.751960 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.757169 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.783994 4969 scope.go:117] "RemoveContainer" containerID="ae892b616ad06e64c04daa6d3f8d132e2326db8c507796a2df01a12da54d9b29" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.805094 4969 scope.go:117] "RemoveContainer" containerID="6149164f902be6d8b639a5f942ae0d60f89cff145bcca988a8150693ef37f0a0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.832084 4969 scope.go:117] "RemoveContainer" containerID="5cd4e588f18cfdc29b3af8abac292c60fe0dff0b4bc314b288658ec204b4747a" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866245 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866290 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866313 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866353 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866494 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866552 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866586 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866618 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866766 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqxww\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-kube-api-access-tqxww\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866823 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.866905 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9d35bef4-01ad-4c3d-b042-9ac2350420a4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.968963 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969059 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969103 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969131 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969157 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969180 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969231 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqxww\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-kube-api-access-tqxww\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969269 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969312 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9d35bef4-01ad-4c3d-b042-9ac2350420a4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969357 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.969383 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.970542 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9d35bef4-01ad-4c3d-b042-9ac2350420a4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.973736 4969 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.973786 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6bc02e5d9caae1a8b7f65724c18b281ae808553b7b037139e0479e345eca79b6/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.974566 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.977172 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.977256 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.977900 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.979684 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.980047 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.980457 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.984647 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:38 crc kubenswrapper[4969]: I1004 08:33:38.992443 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqxww\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-kube-api-access-tqxww\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:39 crc kubenswrapper[4969]: I1004 08:33:39.016616 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:39 crc kubenswrapper[4969]: I1004 08:33:39.055347 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 08:33:39 crc kubenswrapper[4969]: I1004 08:33:39.064959 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de7855b-90f1-46ae-a85c-9e292224b87e" path="/var/lib/kubelet/pods/1de7855b-90f1-46ae-a85c-9e292224b87e/volumes" Oct 04 08:33:39 crc kubenswrapper[4969]: I1004 08:33:39.065632 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49069f68-580b-41f3-9504-6e8b722459f7" path="/var/lib/kubelet/pods/49069f68-580b-41f3-9504-6e8b722459f7/volumes" Oct 04 08:33:45 crc kubenswrapper[4969]: I1004 08:33:45.861722 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.202673 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.269393 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-5bss9"] Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.270514 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5bss9" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.280749 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5bss9"] Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.299306 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7pnm\" (UniqueName: \"kubernetes.io/projected/36354e6c-6917-4239-9ad3-c2cb67c8f765-kube-api-access-r7pnm\") pod \"barbican-db-create-5bss9\" (UID: \"36354e6c-6917-4239-9ad3-c2cb67c8f765\") " pod="openstack/barbican-db-create-5bss9" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.357699 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-9ch7d"] Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.364352 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9ch7d" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.369559 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9ch7d"] Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.400704 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7pnm\" (UniqueName: \"kubernetes.io/projected/36354e6c-6917-4239-9ad3-c2cb67c8f765-kube-api-access-r7pnm\") pod \"barbican-db-create-5bss9\" (UID: \"36354e6c-6917-4239-9ad3-c2cb67c8f765\") " pod="openstack/barbican-db-create-5bss9" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.400791 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmf5z\" (UniqueName: \"kubernetes.io/projected/49667965-1201-4bd7-8b23-adf9f35d866f-kube-api-access-xmf5z\") pod \"cinder-db-create-9ch7d\" (UID: \"49667965-1201-4bd7-8b23-adf9f35d866f\") " pod="openstack/cinder-db-create-9ch7d" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.410573 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-notifications-server-0" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.437389 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7pnm\" (UniqueName: \"kubernetes.io/projected/36354e6c-6917-4239-9ad3-c2cb67c8f765-kube-api-access-r7pnm\") pod \"barbican-db-create-5bss9\" (UID: \"36354e6c-6917-4239-9ad3-c2cb67c8f765\") " pod="openstack/barbican-db-create-5bss9" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.502595 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmf5z\" (UniqueName: \"kubernetes.io/projected/49667965-1201-4bd7-8b23-adf9f35d866f-kube-api-access-xmf5z\") pod \"cinder-db-create-9ch7d\" (UID: \"49667965-1201-4bd7-8b23-adf9f35d866f\") " pod="openstack/cinder-db-create-9ch7d" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.518655 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-7nbv5"] Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.519681 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.525829 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-82wwh" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.525836 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.525981 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.526070 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.537230 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7nbv5"] Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.540276 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmf5z\" (UniqueName: \"kubernetes.io/projected/49667965-1201-4bd7-8b23-adf9f35d866f-kube-api-access-xmf5z\") pod \"cinder-db-create-9ch7d\" (UID: \"49667965-1201-4bd7-8b23-adf9f35d866f\") " pod="openstack/cinder-db-create-9ch7d" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.568883 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-5rtdq"] Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.570194 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5rtdq" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.581577 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5rtdq"] Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.592146 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5bss9" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.605859 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j6tx\" (UniqueName: \"kubernetes.io/projected/550752d6-efab-448d-8776-6f6031942c11-kube-api-access-5j6tx\") pod \"neutron-db-create-5rtdq\" (UID: \"550752d6-efab-448d-8776-6f6031942c11\") " pod="openstack/neutron-db-create-5rtdq" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.605930 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwdhl\" (UniqueName: \"kubernetes.io/projected/bd901023-7f58-4004-bf0e-7df1600efbeb-kube-api-access-pwdhl\") pod \"keystone-db-sync-7nbv5\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.606046 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-combined-ca-bundle\") pod \"keystone-db-sync-7nbv5\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.606139 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-config-data\") pod \"keystone-db-sync-7nbv5\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.679724 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9ch7d" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.707067 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwdhl\" (UniqueName: \"kubernetes.io/projected/bd901023-7f58-4004-bf0e-7df1600efbeb-kube-api-access-pwdhl\") pod \"keystone-db-sync-7nbv5\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.707127 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-combined-ca-bundle\") pod \"keystone-db-sync-7nbv5\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.707173 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-config-data\") pod \"keystone-db-sync-7nbv5\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.707211 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j6tx\" (UniqueName: \"kubernetes.io/projected/550752d6-efab-448d-8776-6f6031942c11-kube-api-access-5j6tx\") pod \"neutron-db-create-5rtdq\" (UID: \"550752d6-efab-448d-8776-6f6031942c11\") " pod="openstack/neutron-db-create-5rtdq" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.711737 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-config-data\") pod \"keystone-db-sync-7nbv5\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.728232 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j6tx\" (UniqueName: \"kubernetes.io/projected/550752d6-efab-448d-8776-6f6031942c11-kube-api-access-5j6tx\") pod \"neutron-db-create-5rtdq\" (UID: \"550752d6-efab-448d-8776-6f6031942c11\") " pod="openstack/neutron-db-create-5rtdq" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.749617 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-combined-ca-bundle\") pod \"keystone-db-sync-7nbv5\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.754004 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwdhl\" (UniqueName: \"kubernetes.io/projected/bd901023-7f58-4004-bf0e-7df1600efbeb-kube-api-access-pwdhl\") pod \"keystone-db-sync-7nbv5\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.890286 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:33:46 crc kubenswrapper[4969]: I1004 08:33:46.899147 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5rtdq" Oct 04 08:33:47 crc kubenswrapper[4969]: I1004 08:33:47.598982 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5bss9"] Oct 04 08:33:47 crc kubenswrapper[4969]: I1004 08:33:47.621870 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9ch7d"] Oct 04 08:33:47 crc kubenswrapper[4969]: W1004 08:33:47.644764 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49667965_1201_4bd7_8b23_adf9f35d866f.slice/crio-a009089a795a1b2ca50fe7e26a4d194e95464d2b4a8e53302884e902f350e5e3 WatchSource:0}: Error finding container a009089a795a1b2ca50fe7e26a4d194e95464d2b4a8e53302884e902f350e5e3: Status 404 returned error can't find the container with id a009089a795a1b2ca50fe7e26a4d194e95464d2b4a8e53302884e902f350e5e3 Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:47.741146 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5bss9" event={"ID":"36354e6c-6917-4239-9ad3-c2cb67c8f765","Type":"ContainerStarted","Data":"8c07e6ae577f100d96616c9a550638d5d0b20e90abb102579d5b5bcd4f45fb9f"} Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:47.751272 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9ch7d" event={"ID":"49667965-1201-4bd7-8b23-adf9f35d866f","Type":"ContainerStarted","Data":"a009089a795a1b2ca50fe7e26a4d194e95464d2b4a8e53302884e902f350e5e3"} Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:47.875523 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7nbv5"] Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:47.890067 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.010233 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5rtdq"] Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.763002 4969 generic.go:334] "Generic (PLEG): container finished" podID="36354e6c-6917-4239-9ad3-c2cb67c8f765" containerID="b1f95ee6222b1871dc4a4a676b443bf3d0733ab09387a6662fd2ea3d4a585122" exitCode=0 Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.763515 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5bss9" event={"ID":"36354e6c-6917-4239-9ad3-c2cb67c8f765","Type":"ContainerDied","Data":"b1f95ee6222b1871dc4a4a676b443bf3d0733ab09387a6662fd2ea3d4a585122"} Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.770592 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7nbv5" event={"ID":"bd901023-7f58-4004-bf0e-7df1600efbeb","Type":"ContainerStarted","Data":"8c93d035647f4706de2cc3cd088747d0cc475eefee9b4f2e1e6e91909d14b5d1"} Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.776579 4969 generic.go:334] "Generic (PLEG): container finished" podID="49667965-1201-4bd7-8b23-adf9f35d866f" containerID="f2a430039a366420a178b21f2dd8fe7f8320fadfcaf782a05c8b1dc2c1d5993e" exitCode=0 Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.776708 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9ch7d" event={"ID":"49667965-1201-4bd7-8b23-adf9f35d866f","Type":"ContainerDied","Data":"f2a430039a366420a178b21f2dd8fe7f8320fadfcaf782a05c8b1dc2c1d5993e"} Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.779340 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2t48v" event={"ID":"ed6a88af-0fc2-45f8-b584-451c87b14ae6","Type":"ContainerStarted","Data":"a5f9b29e226f1f785d18020b72e1529e218c6e168a1c5a730852f368d851c3c4"} Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.781673 4969 generic.go:334] "Generic (PLEG): container finished" podID="550752d6-efab-448d-8776-6f6031942c11" containerID="a7a695ad30ea674847a4264ecfb2baeae12a9c427cd584bc0a8ccc33d1b4d8e8" exitCode=0 Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.781750 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5rtdq" event={"ID":"550752d6-efab-448d-8776-6f6031942c11","Type":"ContainerDied","Data":"a7a695ad30ea674847a4264ecfb2baeae12a9c427cd584bc0a8ccc33d1b4d8e8"} Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.781775 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5rtdq" event={"ID":"550752d6-efab-448d-8776-6f6031942c11","Type":"ContainerStarted","Data":"567e346e85e3167ade660c7a710600b9c94e0650400311a5a47e5de3c3798969"} Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.783190 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerStarted","Data":"4f3c57ab7e7f4de9e58e0516b0c3ec637a1a71eb9a032938c141cf03b1ce54c5"} Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.823179 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-2t48v" podStartSLOduration=3.693313721 podStartE2EDuration="17.823161321s" podCreationTimestamp="2025-10-04 08:33:31 +0000 UTC" firstStartedPulling="2025-10-04 08:33:32.923139184 +0000 UTC m=+1040.677407998" lastFinishedPulling="2025-10-04 08:33:47.052986784 +0000 UTC m=+1054.807255598" observedRunningTime="2025-10-04 08:33:48.819786587 +0000 UTC m=+1056.574055411" watchObservedRunningTime="2025-10-04 08:33:48.823161321 +0000 UTC m=+1056.577430145" Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.896622 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-rskkf"] Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.899943 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.902090 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-szkr4" Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.902412 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Oct 04 08:33:48 crc kubenswrapper[4969]: I1004 08:33:48.915536 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-rskkf"] Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.058316 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-combined-ca-bundle\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.058396 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-config-data\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.058437 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g46hx\" (UniqueName: \"kubernetes.io/projected/35c12eab-f598-4088-b689-ad192cd83887-kube-api-access-g46hx\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.058453 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-db-sync-config-data\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.160547 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-combined-ca-bundle\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.160674 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-config-data\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.160706 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g46hx\" (UniqueName: \"kubernetes.io/projected/35c12eab-f598-4088-b689-ad192cd83887-kube-api-access-g46hx\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.160729 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-db-sync-config-data\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.177262 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-config-data\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.177362 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-combined-ca-bundle\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.178515 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g46hx\" (UniqueName: \"kubernetes.io/projected/35c12eab-f598-4088-b689-ad192cd83887-kube-api-access-g46hx\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.191564 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-db-sync-config-data\") pod \"watcher-db-sync-rskkf\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.217948 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-rskkf" Oct 04 08:33:49 crc kubenswrapper[4969]: I1004 08:33:49.693089 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-rskkf"] Oct 04 08:33:50 crc kubenswrapper[4969]: I1004 08:33:50.802669 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerStarted","Data":"5ae2543215bb1ced07079e957d140539dd1ef5d1a783c89a8c8fe25e15fca67b"} Oct 04 08:33:52 crc kubenswrapper[4969]: W1004 08:33:52.773276 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35c12eab_f598_4088_b689_ad192cd83887.slice/crio-cad6697dcbc70d060504cd3ce8618cf10fa648144f154fe9f8d89491030f05ed WatchSource:0}: Error finding container cad6697dcbc70d060504cd3ce8618cf10fa648144f154fe9f8d89491030f05ed: Status 404 returned error can't find the container with id cad6697dcbc70d060504cd3ce8618cf10fa648144f154fe9f8d89491030f05ed Oct 04 08:33:52 crc kubenswrapper[4969]: I1004 08:33:52.852110 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-rskkf" event={"ID":"35c12eab-f598-4088-b689-ad192cd83887","Type":"ContainerStarted","Data":"cad6697dcbc70d060504cd3ce8618cf10fa648144f154fe9f8d89491030f05ed"} Oct 04 08:33:52 crc kubenswrapper[4969]: I1004 08:33:52.855524 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5rtdq" event={"ID":"550752d6-efab-448d-8776-6f6031942c11","Type":"ContainerDied","Data":"567e346e85e3167ade660c7a710600b9c94e0650400311a5a47e5de3c3798969"} Oct 04 08:33:52 crc kubenswrapper[4969]: I1004 08:33:52.855555 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="567e346e85e3167ade660c7a710600b9c94e0650400311a5a47e5de3c3798969" Oct 04 08:33:52 crc kubenswrapper[4969]: I1004 08:33:52.857080 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5bss9" event={"ID":"36354e6c-6917-4239-9ad3-c2cb67c8f765","Type":"ContainerDied","Data":"8c07e6ae577f100d96616c9a550638d5d0b20e90abb102579d5b5bcd4f45fb9f"} Oct 04 08:33:52 crc kubenswrapper[4969]: I1004 08:33:52.857102 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c07e6ae577f100d96616c9a550638d5d0b20e90abb102579d5b5bcd4f45fb9f" Oct 04 08:33:52 crc kubenswrapper[4969]: I1004 08:33:52.860998 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9ch7d" event={"ID":"49667965-1201-4bd7-8b23-adf9f35d866f","Type":"ContainerDied","Data":"a009089a795a1b2ca50fe7e26a4d194e95464d2b4a8e53302884e902f350e5e3"} Oct 04 08:33:52 crc kubenswrapper[4969]: I1004 08:33:52.861020 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a009089a795a1b2ca50fe7e26a4d194e95464d2b4a8e53302884e902f350e5e3" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.070167 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5bss9" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.090627 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5rtdq" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.092210 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9ch7d" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.259463 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7pnm\" (UniqueName: \"kubernetes.io/projected/36354e6c-6917-4239-9ad3-c2cb67c8f765-kube-api-access-r7pnm\") pod \"36354e6c-6917-4239-9ad3-c2cb67c8f765\" (UID: \"36354e6c-6917-4239-9ad3-c2cb67c8f765\") " Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.259631 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j6tx\" (UniqueName: \"kubernetes.io/projected/550752d6-efab-448d-8776-6f6031942c11-kube-api-access-5j6tx\") pod \"550752d6-efab-448d-8776-6f6031942c11\" (UID: \"550752d6-efab-448d-8776-6f6031942c11\") " Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.259695 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmf5z\" (UniqueName: \"kubernetes.io/projected/49667965-1201-4bd7-8b23-adf9f35d866f-kube-api-access-xmf5z\") pod \"49667965-1201-4bd7-8b23-adf9f35d866f\" (UID: \"49667965-1201-4bd7-8b23-adf9f35d866f\") " Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.264521 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49667965-1201-4bd7-8b23-adf9f35d866f-kube-api-access-xmf5z" (OuterVolumeSpecName: "kube-api-access-xmf5z") pod "49667965-1201-4bd7-8b23-adf9f35d866f" (UID: "49667965-1201-4bd7-8b23-adf9f35d866f"). InnerVolumeSpecName "kube-api-access-xmf5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.264969 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36354e6c-6917-4239-9ad3-c2cb67c8f765-kube-api-access-r7pnm" (OuterVolumeSpecName: "kube-api-access-r7pnm") pod "36354e6c-6917-4239-9ad3-c2cb67c8f765" (UID: "36354e6c-6917-4239-9ad3-c2cb67c8f765"). InnerVolumeSpecName "kube-api-access-r7pnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.265610 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/550752d6-efab-448d-8776-6f6031942c11-kube-api-access-5j6tx" (OuterVolumeSpecName: "kube-api-access-5j6tx") pod "550752d6-efab-448d-8776-6f6031942c11" (UID: "550752d6-efab-448d-8776-6f6031942c11"). InnerVolumeSpecName "kube-api-access-5j6tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.361869 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7pnm\" (UniqueName: \"kubernetes.io/projected/36354e6c-6917-4239-9ad3-c2cb67c8f765-kube-api-access-r7pnm\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.361899 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j6tx\" (UniqueName: \"kubernetes.io/projected/550752d6-efab-448d-8776-6f6031942c11-kube-api-access-5j6tx\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.361908 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmf5z\" (UniqueName: \"kubernetes.io/projected/49667965-1201-4bd7-8b23-adf9f35d866f-kube-api-access-xmf5z\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.875951 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9ch7d" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.875944 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7nbv5" event={"ID":"bd901023-7f58-4004-bf0e-7df1600efbeb","Type":"ContainerStarted","Data":"3e6ffec15040977a3b6cc2dfc19a9cce09ee68281ab8c8e130437cc576655f49"} Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.876020 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5rtdq" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.876063 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5bss9" Oct 04 08:33:53 crc kubenswrapper[4969]: I1004 08:33:53.903257 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-7nbv5" podStartSLOduration=2.947428155 podStartE2EDuration="7.903236725s" podCreationTimestamp="2025-10-04 08:33:46 +0000 UTC" firstStartedPulling="2025-10-04 08:33:47.893490588 +0000 UTC m=+1055.647759402" lastFinishedPulling="2025-10-04 08:33:52.849299138 +0000 UTC m=+1060.603567972" observedRunningTime="2025-10-04 08:33:53.891689398 +0000 UTC m=+1061.645958212" watchObservedRunningTime="2025-10-04 08:33:53.903236725 +0000 UTC m=+1061.657505559" Oct 04 08:33:56 crc kubenswrapper[4969]: I1004 08:33:56.904520 4969 generic.go:334] "Generic (PLEG): container finished" podID="ed6a88af-0fc2-45f8-b584-451c87b14ae6" containerID="a5f9b29e226f1f785d18020b72e1529e218c6e168a1c5a730852f368d851c3c4" exitCode=0 Oct 04 08:33:56 crc kubenswrapper[4969]: I1004 08:33:56.904635 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2t48v" event={"ID":"ed6a88af-0fc2-45f8-b584-451c87b14ae6","Type":"ContainerDied","Data":"a5f9b29e226f1f785d18020b72e1529e218c6e168a1c5a730852f368d851c3c4"} Oct 04 08:33:56 crc kubenswrapper[4969]: I1004 08:33:56.906347 4969 generic.go:334] "Generic (PLEG): container finished" podID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerID="5ae2543215bb1ced07079e957d140539dd1ef5d1a783c89a8c8fe25e15fca67b" exitCode=0 Oct 04 08:33:56 crc kubenswrapper[4969]: I1004 08:33:56.906381 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerDied","Data":"5ae2543215bb1ced07079e957d140539dd1ef5d1a783c89a8c8fe25e15fca67b"} Oct 04 08:33:58 crc kubenswrapper[4969]: I1004 08:33:58.932888 4969 generic.go:334] "Generic (PLEG): container finished" podID="bd901023-7f58-4004-bf0e-7df1600efbeb" containerID="3e6ffec15040977a3b6cc2dfc19a9cce09ee68281ab8c8e130437cc576655f49" exitCode=0 Oct 04 08:33:58 crc kubenswrapper[4969]: I1004 08:33:58.932982 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7nbv5" event={"ID":"bd901023-7f58-4004-bf0e-7df1600efbeb","Type":"ContainerDied","Data":"3e6ffec15040977a3b6cc2dfc19a9cce09ee68281ab8c8e130437cc576655f49"} Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.590656 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.782119 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-db-sync-config-data\") pod \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.782516 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh7fp\" (UniqueName: \"kubernetes.io/projected/ed6a88af-0fc2-45f8-b584-451c87b14ae6-kube-api-access-dh7fp\") pod \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.782567 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-config-data\") pod \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.782607 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-combined-ca-bundle\") pod \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\" (UID: \"ed6a88af-0fc2-45f8-b584-451c87b14ae6\") " Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.808762 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed6a88af-0fc2-45f8-b584-451c87b14ae6-kube-api-access-dh7fp" (OuterVolumeSpecName: "kube-api-access-dh7fp") pod "ed6a88af-0fc2-45f8-b584-451c87b14ae6" (UID: "ed6a88af-0fc2-45f8-b584-451c87b14ae6"). InnerVolumeSpecName "kube-api-access-dh7fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.815596 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ed6a88af-0fc2-45f8-b584-451c87b14ae6" (UID: "ed6a88af-0fc2-45f8-b584-451c87b14ae6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.846585 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed6a88af-0fc2-45f8-b584-451c87b14ae6" (UID: "ed6a88af-0fc2-45f8-b584-451c87b14ae6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.886458 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.886554 4969 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.886564 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh7fp\" (UniqueName: \"kubernetes.io/projected/ed6a88af-0fc2-45f8-b584-451c87b14ae6-kube-api-access-dh7fp\") on node \"crc\" DevicePath \"\"" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.912744 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-config-data" (OuterVolumeSpecName: "config-data") pod "ed6a88af-0fc2-45f8-b584-451c87b14ae6" (UID: "ed6a88af-0fc2-45f8-b584-451c87b14ae6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.948185 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2t48v" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.948329 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2t48v" event={"ID":"ed6a88af-0fc2-45f8-b584-451c87b14ae6","Type":"ContainerDied","Data":"b76536b25ffe2a368483a084c0d10400c4329b23be844cc2145104c92ab61ec2"} Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.948361 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b76536b25ffe2a368483a084c0d10400c4329b23be844cc2145104c92ab61ec2" Oct 04 08:33:59 crc kubenswrapper[4969]: I1004 08:33:59.987706 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6a88af-0fc2-45f8-b584-451c87b14ae6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.339458 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.496339 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-config-data\") pod \"bd901023-7f58-4004-bf0e-7df1600efbeb\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.497623 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwdhl\" (UniqueName: \"kubernetes.io/projected/bd901023-7f58-4004-bf0e-7df1600efbeb-kube-api-access-pwdhl\") pod \"bd901023-7f58-4004-bf0e-7df1600efbeb\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.497667 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-combined-ca-bundle\") pod \"bd901023-7f58-4004-bf0e-7df1600efbeb\" (UID: \"bd901023-7f58-4004-bf0e-7df1600efbeb\") " Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.502662 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd901023-7f58-4004-bf0e-7df1600efbeb-kube-api-access-pwdhl" (OuterVolumeSpecName: "kube-api-access-pwdhl") pod "bd901023-7f58-4004-bf0e-7df1600efbeb" (UID: "bd901023-7f58-4004-bf0e-7df1600efbeb"). InnerVolumeSpecName "kube-api-access-pwdhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.527573 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd901023-7f58-4004-bf0e-7df1600efbeb" (UID: "bd901023-7f58-4004-bf0e-7df1600efbeb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.548683 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-config-data" (OuterVolumeSpecName: "config-data") pod "bd901023-7f58-4004-bf0e-7df1600efbeb" (UID: "bd901023-7f58-4004-bf0e-7df1600efbeb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.599470 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.599498 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwdhl\" (UniqueName: \"kubernetes.io/projected/bd901023-7f58-4004-bf0e-7df1600efbeb-kube-api-access-pwdhl\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.599509 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd901023-7f58-4004-bf0e-7df1600efbeb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.959391 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7nbv5" event={"ID":"bd901023-7f58-4004-bf0e-7df1600efbeb","Type":"ContainerDied","Data":"8c93d035647f4706de2cc3cd088747d0cc475eefee9b4f2e1e6e91909d14b5d1"} Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.959727 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c93d035647f4706de2cc3cd088747d0cc475eefee9b4f2e1e6e91909d14b5d1" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.959627 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7nbv5" Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.961337 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-rskkf" event={"ID":"35c12eab-f598-4088-b689-ad192cd83887","Type":"ContainerStarted","Data":"b81188eb7053c4f9e2356cb249f9e3eefc37024c49e128545314631891193999"} Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.963451 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerStarted","Data":"7596bb8d092712e53e08b6ed3fc75137915d4fbf69e7f3231d3107c311f09e9b"} Oct 04 08:34:00 crc kubenswrapper[4969]: I1004 08:34:00.987848 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-rskkf" podStartSLOduration=5.347508641 podStartE2EDuration="12.98781197s" podCreationTimestamp="2025-10-04 08:33:48 +0000 UTC" firstStartedPulling="2025-10-04 08:33:52.776795443 +0000 UTC m=+1060.531064257" lastFinishedPulling="2025-10-04 08:34:00.417098772 +0000 UTC m=+1068.171367586" observedRunningTime="2025-10-04 08:34:00.984502057 +0000 UTC m=+1068.738770891" watchObservedRunningTime="2025-10-04 08:34:00.98781197 +0000 UTC m=+1068.742080784" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.047484 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c6d9f9b87-226hp"] Oct 04 08:34:01 crc kubenswrapper[4969]: E1004 08:34:01.047958 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed6a88af-0fc2-45f8-b584-451c87b14ae6" containerName="glance-db-sync" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.047981 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed6a88af-0fc2-45f8-b584-451c87b14ae6" containerName="glance-db-sync" Oct 04 08:34:01 crc kubenswrapper[4969]: E1004 08:34:01.047997 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49667965-1201-4bd7-8b23-adf9f35d866f" containerName="mariadb-database-create" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.048006 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="49667965-1201-4bd7-8b23-adf9f35d866f" containerName="mariadb-database-create" Oct 04 08:34:01 crc kubenswrapper[4969]: E1004 08:34:01.048022 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="550752d6-efab-448d-8776-6f6031942c11" containerName="mariadb-database-create" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.048030 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="550752d6-efab-448d-8776-6f6031942c11" containerName="mariadb-database-create" Oct 04 08:34:01 crc kubenswrapper[4969]: E1004 08:34:01.048056 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd901023-7f58-4004-bf0e-7df1600efbeb" containerName="keystone-db-sync" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.048064 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd901023-7f58-4004-bf0e-7df1600efbeb" containerName="keystone-db-sync" Oct 04 08:34:01 crc kubenswrapper[4969]: E1004 08:34:01.048085 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36354e6c-6917-4239-9ad3-c2cb67c8f765" containerName="mariadb-database-create" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.048093 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="36354e6c-6917-4239-9ad3-c2cb67c8f765" containerName="mariadb-database-create" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.048322 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="49667965-1201-4bd7-8b23-adf9f35d866f" containerName="mariadb-database-create" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.048352 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="550752d6-efab-448d-8776-6f6031942c11" containerName="mariadb-database-create" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.048369 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd901023-7f58-4004-bf0e-7df1600efbeb" containerName="keystone-db-sync" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.048379 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed6a88af-0fc2-45f8-b584-451c87b14ae6" containerName="glance-db-sync" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.048397 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="36354e6c-6917-4239-9ad3-c2cb67c8f765" containerName="mariadb-database-create" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.049516 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.089275 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c6d9f9b87-226hp"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.211470 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-sb\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.211530 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-swift-storage-0\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.211842 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-config\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.211953 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5b5z\" (UniqueName: \"kubernetes.io/projected/ebea1020-49c7-4d07-b289-08255a7b3ad9-kube-api-access-v5b5z\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.212018 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-svc\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.212041 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-nb\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.221789 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fkm5h"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.222984 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.225039 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.225254 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-82wwh" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.225371 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.225583 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.233976 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c6d9f9b87-226hp"] Oct 04 08:34:01 crc kubenswrapper[4969]: E1004 08:34:01.235002 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-v5b5z ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" podUID="ebea1020-49c7-4d07-b289-08255a7b3ad9" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.263097 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fkm5h"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.289021 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869c84485f-4mr79"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.290549 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315610 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-config\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315646 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-fernet-keys\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315669 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6bfr\" (UniqueName: \"kubernetes.io/projected/6e8240c5-920e-4cbd-9028-46f47a239f45-kube-api-access-r6bfr\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315692 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5b5z\" (UniqueName: \"kubernetes.io/projected/ebea1020-49c7-4d07-b289-08255a7b3ad9-kube-api-access-v5b5z\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315720 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-svc\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315735 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-nb\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315760 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-scripts\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315783 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-config-data\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315803 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-credential-keys\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315827 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-sb\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315863 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-swift-storage-0\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.315908 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-combined-ca-bundle\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.316706 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-config\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.317477 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-svc\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.317861 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869c84485f-4mr79"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.317992 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-nb\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.318254 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-sb\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.318818 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-swift-storage-0\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.347410 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5b5z\" (UniqueName: \"kubernetes.io/projected/ebea1020-49c7-4d07-b289-08255a7b3ad9-kube-api-access-v5b5z\") pod \"dnsmasq-dns-5c6d9f9b87-226hp\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418441 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-swift-storage-0\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418510 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-sb\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418544 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkf5m\" (UniqueName: \"kubernetes.io/projected/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-kube-api-access-kkf5m\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418571 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-nb\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418599 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-combined-ca-bundle\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418634 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-config\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418663 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-fernet-keys\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418689 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6bfr\" (UniqueName: \"kubernetes.io/projected/6e8240c5-920e-4cbd-9028-46f47a239f45-kube-api-access-r6bfr\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418740 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-scripts\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418767 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-config-data\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418788 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-svc\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.418813 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-credential-keys\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.422205 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-combined-ca-bundle\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.425133 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-config-data\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.428057 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-credential-keys\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.431538 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-fernet-keys\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.451955 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-59494b9bd5-6j6nz"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.452459 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-scripts\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.473446 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.474031 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6bfr\" (UniqueName: \"kubernetes.io/projected/6e8240c5-920e-4cbd-9028-46f47a239f45-kube-api-access-r6bfr\") pod \"keystone-bootstrap-fkm5h\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.481752 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59494b9bd5-6j6nz"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.484215 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-grsq5" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.484624 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.484810 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.485722 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.521783 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkf5m\" (UniqueName: \"kubernetes.io/projected/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-kube-api-access-kkf5m\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.521830 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-nb\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.521880 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-config\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.521954 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-svc\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.521991 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-sb\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.522008 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-swift-storage-0\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.522825 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-swift-storage-0\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.528980 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-sb\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.529507 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-svc\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.537074 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-config\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.544383 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-nb\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.554468 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.556369 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.562243 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.562327 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.562638 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.586343 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.593633 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkf5m\" (UniqueName: \"kubernetes.io/projected/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-kube-api-access-kkf5m\") pod \"dnsmasq-dns-869c84485f-4mr79\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.610492 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nx8dt"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.611958 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.623987 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plcfr\" (UniqueName: \"kubernetes.io/projected/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-kube-api-access-plcfr\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.624041 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-config-data\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.624067 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-horizon-secret-key\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.624143 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-logs\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.624184 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-scripts\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.624610 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.624820 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.624975 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ssb6j" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.639013 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.694486 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nx8dt"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727501 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-config-data\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727568 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-scripts\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727594 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-run-httpd\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727622 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-config-data\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727650 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w6w9\" (UniqueName: \"kubernetes.io/projected/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-kube-api-access-8w6w9\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727677 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-log-httpd\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727733 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727854 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-scripts\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727895 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plcfr\" (UniqueName: \"kubernetes.io/projected/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-kube-api-access-plcfr\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727921 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-scripts\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727946 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-combined-ca-bundle\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.727985 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-config-data\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.728010 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-horizon-secret-key\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.728060 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.728090 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a60a257-4654-4bfd-b878-94733004ac3d-logs\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.728118 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfl4h\" (UniqueName: \"kubernetes.io/projected/4a60a257-4654-4bfd-b878-94733004ac3d-kube-api-access-bfl4h\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.728181 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-logs\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.728711 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-logs\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.729395 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-scripts\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.731319 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-config-data\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.732534 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869c84485f-4mr79"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.761889 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-horizon-secret-key\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.761980 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79cb9c979c-pr7fn"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.763691 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.776941 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79cb9c979c-pr7fn"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.784059 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plcfr\" (UniqueName: \"kubernetes.io/projected/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-kube-api-access-plcfr\") pod \"horizon-59494b9bd5-6j6nz\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.811461 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b6dcf6db5-m8jj5"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.813028 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.821732 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.823181 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.826469 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vmpm7" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.826761 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833060 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833079 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-logs\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833135 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-horizon-secret-key\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833183 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpfwq\" (UniqueName: \"kubernetes.io/projected/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-kube-api-access-bpfwq\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833223 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-scripts\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833244 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-config-data\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833277 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-scripts\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833523 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-combined-ca-bundle\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833568 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-scripts\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833607 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833644 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a60a257-4654-4bfd-b878-94733004ac3d-logs\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833673 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfl4h\" (UniqueName: \"kubernetes.io/projected/4a60a257-4654-4bfd-b878-94733004ac3d-kube-api-access-bfl4h\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833746 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-config-data\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833773 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-run-httpd\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833799 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-config-data\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833824 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w6w9\" (UniqueName: \"kubernetes.io/projected/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-kube-api-access-8w6w9\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833850 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-log-httpd\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.833897 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.834858 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a60a257-4654-4bfd-b878-94733004ac3d-logs\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.841924 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-run-httpd\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.843506 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-log-httpd\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.853830 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-scripts\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.855138 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b6dcf6db5-m8jj5"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.855285 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-config-data\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.855866 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.855950 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-config-data\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.860058 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.865145 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.865890 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.868617 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-combined-ca-bundle\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.869702 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-scripts\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.874897 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w6w9\" (UniqueName: \"kubernetes.io/projected/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-kube-api-access-8w6w9\") pod \"ceilometer-0\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " pod="openstack/ceilometer-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.878960 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfl4h\" (UniqueName: \"kubernetes.io/projected/4a60a257-4654-4bfd-b878-94733004ac3d-kube-api-access-bfl4h\") pod \"placement-db-sync-nx8dt\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.881610 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.890291 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.893550 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.897299 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935014 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-svc\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935071 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935177 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-logs\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935204 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cc76\" (UniqueName: \"kubernetes.io/projected/d7d78c6c-c9e9-4a37-babc-801522168a84-kube-api-access-5cc76\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935229 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-horizon-secret-key\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935247 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4vc7\" (UniqueName: \"kubernetes.io/projected/9341083c-6983-497b-a0a7-fcca75c3ba65-kube-api-access-x4vc7\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935273 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpfwq\" (UniqueName: \"kubernetes.io/projected/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-kube-api-access-bpfwq\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935298 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-config-data\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935327 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-nb\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935347 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-sb\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935365 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-scripts\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935386 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-config-data\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935404 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935442 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-logs\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935476 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-swift-storage-0\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935501 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-config\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935521 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-scripts\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935540 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.935964 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-logs\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.937119 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-scripts\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.939617 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-config-data\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.946682 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-horizon-secret-key\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:01 crc kubenswrapper[4969]: I1004 08:34:01.956109 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.002480 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.019768 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.036646 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042456 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-swift-storage-0\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042514 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kpt2\" (UniqueName: \"kubernetes.io/projected/1e850dc8-3f10-4136-b3e6-8b1c13809888-kube-api-access-5kpt2\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042622 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-config\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042668 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-scripts\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042730 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042762 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042786 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-svc\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042843 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042903 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-logs\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.042937 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cc76\" (UniqueName: \"kubernetes.io/projected/d7d78c6c-c9e9-4a37-babc-801522168a84-kube-api-access-5cc76\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043002 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4vc7\" (UniqueName: \"kubernetes.io/projected/9341083c-6983-497b-a0a7-fcca75c3ba65-kube-api-access-x4vc7\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043035 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043247 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-nb\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043280 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043315 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-sb\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043380 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-config-data\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043412 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043543 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043598 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.043635 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-logs\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.044240 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-logs\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.044645 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-config\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.048757 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.053166 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.053923 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpfwq\" (UniqueName: \"kubernetes.io/projected/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-kube-api-access-bpfwq\") pod \"horizon-79cb9c979c-pr7fn\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.055755 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-swift-storage-0\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.056407 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-nb\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.056526 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-sb\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.057000 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-svc\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.080152 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-config-data\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.119238 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-scripts\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.132239 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4vc7\" (UniqueName: \"kubernetes.io/projected/9341083c-6983-497b-a0a7-fcca75c3ba65-kube-api-access-x4vc7\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.132252 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cc76\" (UniqueName: \"kubernetes.io/projected/d7d78c6c-c9e9-4a37-babc-801522168a84-kube-api-access-5cc76\") pod \"dnsmasq-dns-7b6dcf6db5-m8jj5\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.151117 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-sb\") pod \"ebea1020-49c7-4d07-b289-08255a7b3ad9\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.151221 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5b5z\" (UniqueName: \"kubernetes.io/projected/ebea1020-49c7-4d07-b289-08255a7b3ad9-kube-api-access-v5b5z\") pod \"ebea1020-49c7-4d07-b289-08255a7b3ad9\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.151256 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-swift-storage-0\") pod \"ebea1020-49c7-4d07-b289-08255a7b3ad9\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.151731 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-nb\") pod \"ebea1020-49c7-4d07-b289-08255a7b3ad9\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.151767 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-svc\") pod \"ebea1020-49c7-4d07-b289-08255a7b3ad9\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.151815 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-config\") pod \"ebea1020-49c7-4d07-b289-08255a7b3ad9\" (UID: \"ebea1020-49c7-4d07-b289-08255a7b3ad9\") " Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.152499 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kpt2\" (UniqueName: \"kubernetes.io/projected/1e850dc8-3f10-4136-b3e6-8b1c13809888-kube-api-access-5kpt2\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.152586 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.152647 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-logs\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.152690 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.152756 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.152805 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.152831 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.153404 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.159397 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.166974 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ebea1020-49c7-4d07-b289-08255a7b3ad9" (UID: "ebea1020-49c7-4d07-b289-08255a7b3ad9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.167512 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ebea1020-49c7-4d07-b289-08255a7b3ad9" (UID: "ebea1020-49c7-4d07-b289-08255a7b3ad9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.167925 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-config" (OuterVolumeSpecName: "config") pod "ebea1020-49c7-4d07-b289-08255a7b3ad9" (UID: "ebea1020-49c7-4d07-b289-08255a7b3ad9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.169771 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ebea1020-49c7-4d07-b289-08255a7b3ad9" (UID: "ebea1020-49c7-4d07-b289-08255a7b3ad9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.191721 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebea1020-49c7-4d07-b289-08255a7b3ad9-kube-api-access-v5b5z" (OuterVolumeSpecName: "kube-api-access-v5b5z") pod "ebea1020-49c7-4d07-b289-08255a7b3ad9" (UID: "ebea1020-49c7-4d07-b289-08255a7b3ad9"). InnerVolumeSpecName "kube-api-access-v5b5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.191823 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.191956 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-logs\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.192323 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ebea1020-49c7-4d07-b289-08255a7b3ad9" (UID: "ebea1020-49c7-4d07-b289-08255a7b3ad9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.193305 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.195863 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.199491 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kpt2\" (UniqueName: \"kubernetes.io/projected/1e850dc8-3f10-4136-b3e6-8b1c13809888-kube-api-access-5kpt2\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.202193 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.219180 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.256804 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.256836 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5b5z\" (UniqueName: \"kubernetes.io/projected/ebea1020-49c7-4d07-b289-08255a7b3ad9-kube-api-access-v5b5z\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.256845 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.256853 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.256863 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.256870 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebea1020-49c7-4d07-b289-08255a7b3ad9-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.259062 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.295659 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.385602 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.412325 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869c84485f-4mr79"] Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.575118 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.599091 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.621025 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fkm5h"] Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.877561 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59494b9bd5-6j6nz"] Oct 04 08:34:02 crc kubenswrapper[4969]: I1004 08:34:02.936314 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.017524 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fkm5h" event={"ID":"6e8240c5-920e-4cbd-9028-46f47a239f45","Type":"ContainerStarted","Data":"6008a5d1a943387685fb479d8c2ef8bf6344c82ccc4309e74bf45684bb2becac"} Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.018718 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59494b9bd5-6j6nz" event={"ID":"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32","Type":"ContainerStarted","Data":"c3a6b6a3c39f0040245b522d336b6e0e9993fc1761f15e1c0510ddd8f9044bb6"} Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.019564 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869c84485f-4mr79" event={"ID":"c07fbf85-1e7e-45e0-bbad-3a1682ed0410","Type":"ContainerStarted","Data":"4603f2b7f4fab2a9ccbc722a634ab2993b24a2dbb412f96c83e83af91f153623"} Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.020474 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6d9f9b87-226hp" Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.021600 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff1b1453-89a7-4b5e-a55d-246e10fa18c2","Type":"ContainerStarted","Data":"8eb03613ffbe4b3fc758b52cd1d456d72cfc2842c4467052b751aebbfa1e0318"} Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.095985 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c6d9f9b87-226hp"] Oct 04 08:34:03 crc kubenswrapper[4969]: W1004 08:34:03.104474 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a60a257_4654_4bfd_b878_94733004ac3d.slice/crio-a22ce1ba338ee94f01c93daa4d6480c09adcc077b78c695a9ec5e1cfeaa1a58a WatchSource:0}: Error finding container a22ce1ba338ee94f01c93daa4d6480c09adcc077b78c695a9ec5e1cfeaa1a58a: Status 404 returned error can't find the container with id a22ce1ba338ee94f01c93daa4d6480c09adcc077b78c695a9ec5e1cfeaa1a58a Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.109045 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nx8dt"] Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.118900 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c6d9f9b87-226hp"] Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.252122 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79cb9c979c-pr7fn"] Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.304557 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b6dcf6db5-m8jj5"] Oct 04 08:34:03 crc kubenswrapper[4969]: W1004 08:34:03.308082 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7d78c6c_c9e9_4a37_babc_801522168a84.slice/crio-f52f38b3698b990789778ad1807ee76316451a626193cb0124c746c64450c32b WatchSource:0}: Error finding container f52f38b3698b990789778ad1807ee76316451a626193cb0124c746c64450c32b: Status 404 returned error can't find the container with id f52f38b3698b990789778ad1807ee76316451a626193cb0124c746c64450c32b Oct 04 08:34:03 crc kubenswrapper[4969]: I1004 08:34:03.590120 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:34:03 crc kubenswrapper[4969]: W1004 08:34:03.612008 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e850dc8_3f10_4136_b3e6_8b1c13809888.slice/crio-43ef672942d27c5294361aacee95297a94289e1d99380c5f68f157c3da7ae115 WatchSource:0}: Error finding container 43ef672942d27c5294361aacee95297a94289e1d99380c5f68f157c3da7ae115: Status 404 returned error can't find the container with id 43ef672942d27c5294361aacee95297a94289e1d99380c5f68f157c3da7ae115 Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.031064 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" event={"ID":"d7d78c6c-c9e9-4a37-babc-801522168a84","Type":"ContainerStarted","Data":"9b070738e94db6fa89d7792e706587342acd0ab263810ad74ec784aca57447b0"} Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.031115 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" event={"ID":"d7d78c6c-c9e9-4a37-babc-801522168a84","Type":"ContainerStarted","Data":"f52f38b3698b990789778ad1807ee76316451a626193cb0124c746c64450c32b"} Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.032749 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nx8dt" event={"ID":"4a60a257-4654-4bfd-b878-94733004ac3d","Type":"ContainerStarted","Data":"a22ce1ba338ee94f01c93daa4d6480c09adcc077b78c695a9ec5e1cfeaa1a58a"} Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.036344 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fkm5h" event={"ID":"6e8240c5-920e-4cbd-9028-46f47a239f45","Type":"ContainerStarted","Data":"b7e88c823b7a23fb328c159dd65d4e4806e654ea40682085f11dd7ccb3d94b19"} Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.038101 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e850dc8-3f10-4136-b3e6-8b1c13809888","Type":"ContainerStarted","Data":"43ef672942d27c5294361aacee95297a94289e1d99380c5f68f157c3da7ae115"} Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.042983 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerStarted","Data":"5f25df52f533494ca9e9131af20ae1e34564c497a31d70b745543bf8fc89fb56"} Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.043018 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerStarted","Data":"fbb54ba479493c3fb0ddbfd2ac2e97371dcae3e39fb3de63e6e2dc2f26e1f759"} Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.055697 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.067462 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cb9c979c-pr7fn" event={"ID":"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49","Type":"ContainerStarted","Data":"d948976c7d1b55c4fd8e95ba1b067e17e519e05a0e00d918a56abd4fb5b21c86"} Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.076863 4969 generic.go:334] "Generic (PLEG): container finished" podID="c07fbf85-1e7e-45e0-bbad-3a1682ed0410" containerID="97ae5ad0c733da372fc170fb8390c1c389f4c84644accc1cce90892c6556e11e" exitCode=0 Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.077383 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869c84485f-4mr79" event={"ID":"c07fbf85-1e7e-45e0-bbad-3a1682ed0410","Type":"ContainerDied","Data":"97ae5ad0c733da372fc170fb8390c1c389f4c84644accc1cce90892c6556e11e"} Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.086668 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fkm5h" podStartSLOduration=3.086645362 podStartE2EDuration="3.086645362s" podCreationTimestamp="2025-10-04 08:34:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:04.067504985 +0000 UTC m=+1071.821773809" watchObservedRunningTime="2025-10-04 08:34:04.086645362 +0000 UTC m=+1071.840914186" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.101287 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=26.101270416 podStartE2EDuration="26.101270416s" podCreationTimestamp="2025-10-04 08:33:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:04.094769395 +0000 UTC m=+1071.849038209" watchObservedRunningTime="2025-10-04 08:34:04.101270416 +0000 UTC m=+1071.855539230" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.631066 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.638961 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.735941 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-nb\") pod \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.736227 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-swift-storage-0\") pod \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.736280 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-config\") pod \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.736312 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkf5m\" (UniqueName: \"kubernetes.io/projected/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-kube-api-access-kkf5m\") pod \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.736371 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-sb\") pod \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.736459 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-svc\") pod \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\" (UID: \"c07fbf85-1e7e-45e0-bbad-3a1682ed0410\") " Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.773603 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-kube-api-access-kkf5m" (OuterVolumeSpecName: "kube-api-access-kkf5m") pod "c07fbf85-1e7e-45e0-bbad-3a1682ed0410" (UID: "c07fbf85-1e7e-45e0-bbad-3a1682ed0410"). InnerVolumeSpecName "kube-api-access-kkf5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.798057 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c07fbf85-1e7e-45e0-bbad-3a1682ed0410" (UID: "c07fbf85-1e7e-45e0-bbad-3a1682ed0410"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.804963 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c07fbf85-1e7e-45e0-bbad-3a1682ed0410" (UID: "c07fbf85-1e7e-45e0-bbad-3a1682ed0410"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.805665 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-config" (OuterVolumeSpecName: "config") pod "c07fbf85-1e7e-45e0-bbad-3a1682ed0410" (UID: "c07fbf85-1e7e-45e0-bbad-3a1682ed0410"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.824412 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c07fbf85-1e7e-45e0-bbad-3a1682ed0410" (UID: "c07fbf85-1e7e-45e0-bbad-3a1682ed0410"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.824590 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c07fbf85-1e7e-45e0-bbad-3a1682ed0410" (UID: "c07fbf85-1e7e-45e0-bbad-3a1682ed0410"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.840991 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.841035 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.841051 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.841065 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.841076 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkf5m\" (UniqueName: \"kubernetes.io/projected/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-kube-api-access-kkf5m\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:04 crc kubenswrapper[4969]: I1004 08:34:04.841087 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c07fbf85-1e7e-45e0-bbad-3a1682ed0410-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.068014 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebea1020-49c7-4d07-b289-08255a7b3ad9" path="/var/lib/kubelet/pods/ebea1020-49c7-4d07-b289-08255a7b3ad9/volumes" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.108661 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e850dc8-3f10-4136-b3e6-8b1c13809888","Type":"ContainerStarted","Data":"3d959c8a63aa449f5dc6a40de75353068c1b01e47255024fb3e4b96e593aa1fa"} Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.113336 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869c84485f-4mr79" event={"ID":"c07fbf85-1e7e-45e0-bbad-3a1682ed0410","Type":"ContainerDied","Data":"4603f2b7f4fab2a9ccbc722a634ab2993b24a2dbb412f96c83e83af91f153623"} Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.113686 4969 scope.go:117] "RemoveContainer" containerID="97ae5ad0c733da372fc170fb8390c1c389f4c84644accc1cce90892c6556e11e" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.113348 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869c84485f-4mr79" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.115558 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9341083c-6983-497b-a0a7-fcca75c3ba65","Type":"ContainerStarted","Data":"a618fb301b028d4978d5c15802cab6df3fbb10b7618022bde0611f15646cd5ae"} Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.140457 4969 generic.go:334] "Generic (PLEG): container finished" podID="d7d78c6c-c9e9-4a37-babc-801522168a84" containerID="9b070738e94db6fa89d7792e706587342acd0ab263810ad74ec784aca57447b0" exitCode=0 Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.140698 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" event={"ID":"d7d78c6c-c9e9-4a37-babc-801522168a84","Type":"ContainerDied","Data":"9b070738e94db6fa89d7792e706587342acd0ab263810ad74ec784aca57447b0"} Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.179249 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869c84485f-4mr79"] Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.201875 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869c84485f-4mr79"] Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.451719 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.472724 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59494b9bd5-6j6nz"] Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.492286 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.525190 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c876d6d6f-vshkz"] Oct 04 08:34:05 crc kubenswrapper[4969]: E1004 08:34:05.525738 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c07fbf85-1e7e-45e0-bbad-3a1682ed0410" containerName="init" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.525760 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c07fbf85-1e7e-45e0-bbad-3a1682ed0410" containerName="init" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.525992 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c07fbf85-1e7e-45e0-bbad-3a1682ed0410" containerName="init" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.527297 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.530945 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c876d6d6f-vshkz"] Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.566795 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-scripts\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.566848 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-logs\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.566891 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbj29\" (UniqueName: \"kubernetes.io/projected/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-kube-api-access-wbj29\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.566918 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-horizon-secret-key\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.566999 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-config-data\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.593267 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.669194 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-scripts\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.669273 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-logs\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.669380 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbj29\" (UniqueName: \"kubernetes.io/projected/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-kube-api-access-wbj29\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.669437 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-horizon-secret-key\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.669678 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-config-data\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.670599 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-logs\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.672151 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-config-data\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.672836 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-scripts\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.676379 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-horizon-secret-key\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.702099 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbj29\" (UniqueName: \"kubernetes.io/projected/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-kube-api-access-wbj29\") pod \"horizon-c876d6d6f-vshkz\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:05 crc kubenswrapper[4969]: I1004 08:34:05.936835 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.154088 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9341083c-6983-497b-a0a7-fcca75c3ba65","Type":"ContainerStarted","Data":"6a8d2051dd19e9a1d1e84a983702d898b445a62dbb102288d6ee49faf5680a83"} Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.155890 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" event={"ID":"d7d78c6c-c9e9-4a37-babc-801522168a84","Type":"ContainerStarted","Data":"c6d8e4496044b41533943667c339f923c40c096a7ecc3cdb6a3dd26e0656626f"} Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.156231 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.195496 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" podStartSLOduration=5.19547448 podStartE2EDuration="5.19547448s" podCreationTimestamp="2025-10-04 08:34:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:06.179527793 +0000 UTC m=+1073.933796597" watchObservedRunningTime="2025-10-04 08:34:06.19547448 +0000 UTC m=+1073.949743304" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.338604 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-0dbb-account-create-k45jh"] Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.339826 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0dbb-account-create-k45jh" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.342666 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.348120 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0dbb-account-create-k45jh"] Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.406277 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lhh4\" (UniqueName: \"kubernetes.io/projected/ef5c63b3-e8f7-4b9a-bb7c-642516049ee5-kube-api-access-2lhh4\") pod \"barbican-0dbb-account-create-k45jh\" (UID: \"ef5c63b3-e8f7-4b9a-bb7c-642516049ee5\") " pod="openstack/barbican-0dbb-account-create-k45jh" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.474929 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c876d6d6f-vshkz"] Oct 04 08:34:06 crc kubenswrapper[4969]: W1004 08:34:06.485058 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf3fb747_e585_4f1f_b0b5_b7fa3ef30f75.slice/crio-fbd9f40bbfa205a8b1d61e8a026306b242faaab6a69b934bfba9ebf1e4be56bf WatchSource:0}: Error finding container fbd9f40bbfa205a8b1d61e8a026306b242faaab6a69b934bfba9ebf1e4be56bf: Status 404 returned error can't find the container with id fbd9f40bbfa205a8b1d61e8a026306b242faaab6a69b934bfba9ebf1e4be56bf Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.508723 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lhh4\" (UniqueName: \"kubernetes.io/projected/ef5c63b3-e8f7-4b9a-bb7c-642516049ee5-kube-api-access-2lhh4\") pod \"barbican-0dbb-account-create-k45jh\" (UID: \"ef5c63b3-e8f7-4b9a-bb7c-642516049ee5\") " pod="openstack/barbican-0dbb-account-create-k45jh" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.544263 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lhh4\" (UniqueName: \"kubernetes.io/projected/ef5c63b3-e8f7-4b9a-bb7c-642516049ee5-kube-api-access-2lhh4\") pod \"barbican-0dbb-account-create-k45jh\" (UID: \"ef5c63b3-e8f7-4b9a-bb7c-642516049ee5\") " pod="openstack/barbican-0dbb-account-create-k45jh" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.549152 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d9f0-account-create-5zlxv"] Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.552709 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d9f0-account-create-5zlxv" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.557339 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.581038 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d9f0-account-create-5zlxv"] Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.611547 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lkjh\" (UniqueName: \"kubernetes.io/projected/93ccd81b-783b-4c03-b564-425601b5df4c-kube-api-access-7lkjh\") pod \"cinder-d9f0-account-create-5zlxv\" (UID: \"93ccd81b-783b-4c03-b564-425601b5df4c\") " pod="openstack/cinder-d9f0-account-create-5zlxv" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.718488 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lkjh\" (UniqueName: \"kubernetes.io/projected/93ccd81b-783b-4c03-b564-425601b5df4c-kube-api-access-7lkjh\") pod \"cinder-d9f0-account-create-5zlxv\" (UID: \"93ccd81b-783b-4c03-b564-425601b5df4c\") " pod="openstack/cinder-d9f0-account-create-5zlxv" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.725911 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0dbb-account-create-k45jh" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.727269 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6ab2-account-create-mpxpw"] Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.728560 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ab2-account-create-mpxpw" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.730949 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.737851 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6ab2-account-create-mpxpw"] Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.751258 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lkjh\" (UniqueName: \"kubernetes.io/projected/93ccd81b-783b-4c03-b564-425601b5df4c-kube-api-access-7lkjh\") pod \"cinder-d9f0-account-create-5zlxv\" (UID: \"93ccd81b-783b-4c03-b564-425601b5df4c\") " pod="openstack/cinder-d9f0-account-create-5zlxv" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.823705 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxcbc\" (UniqueName: \"kubernetes.io/projected/a873063d-e103-48c3-b85e-1b8b2e0ba249-kube-api-access-gxcbc\") pod \"neutron-6ab2-account-create-mpxpw\" (UID: \"a873063d-e103-48c3-b85e-1b8b2e0ba249\") " pod="openstack/neutron-6ab2-account-create-mpxpw" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.927096 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxcbc\" (UniqueName: \"kubernetes.io/projected/a873063d-e103-48c3-b85e-1b8b2e0ba249-kube-api-access-gxcbc\") pod \"neutron-6ab2-account-create-mpxpw\" (UID: \"a873063d-e103-48c3-b85e-1b8b2e0ba249\") " pod="openstack/neutron-6ab2-account-create-mpxpw" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.945880 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d9f0-account-create-5zlxv" Oct 04 08:34:06 crc kubenswrapper[4969]: I1004 08:34:06.986385 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxcbc\" (UniqueName: \"kubernetes.io/projected/a873063d-e103-48c3-b85e-1b8b2e0ba249-kube-api-access-gxcbc\") pod \"neutron-6ab2-account-create-mpxpw\" (UID: \"a873063d-e103-48c3-b85e-1b8b2e0ba249\") " pod="openstack/neutron-6ab2-account-create-mpxpw" Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.051800 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ab2-account-create-mpxpw" Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.076693 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c07fbf85-1e7e-45e0-bbad-3a1682ed0410" path="/var/lib/kubelet/pods/c07fbf85-1e7e-45e0-bbad-3a1682ed0410/volumes" Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.203633 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c876d6d6f-vshkz" event={"ID":"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75","Type":"ContainerStarted","Data":"fbd9f40bbfa205a8b1d61e8a026306b242faaab6a69b934bfba9ebf1e4be56bf"} Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.224189 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerName="glance-log" containerID="cri-o://3d959c8a63aa449f5dc6a40de75353068c1b01e47255024fb3e4b96e593aa1fa" gracePeriod=30 Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.224519 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e850dc8-3f10-4136-b3e6-8b1c13809888","Type":"ContainerStarted","Data":"51ef310c1ede4681abb76c58539854417a7ed1b944e97e43716b31a8d0482342"} Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.224813 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerName="glance-httpd" containerID="cri-o://51ef310c1ede4681abb76c58539854417a7ed1b944e97e43716b31a8d0482342" gracePeriod=30 Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.281063 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.281041263 podStartE2EDuration="6.281041263s" podCreationTimestamp="2025-10-04 08:34:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:07.241935171 +0000 UTC m=+1074.996203985" watchObservedRunningTime="2025-10-04 08:34:07.281041263 +0000 UTC m=+1075.035310077" Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.573022 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d9f0-account-create-5zlxv"] Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.660025 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0dbb-account-create-k45jh"] Oct 04 08:34:07 crc kubenswrapper[4969]: I1004 08:34:07.738965 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6ab2-account-create-mpxpw"] Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.236070 4969 generic.go:334] "Generic (PLEG): container finished" podID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerID="51ef310c1ede4681abb76c58539854417a7ed1b944e97e43716b31a8d0482342" exitCode=0 Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.236095 4969 generic.go:334] "Generic (PLEG): container finished" podID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerID="3d959c8a63aa449f5dc6a40de75353068c1b01e47255024fb3e4b96e593aa1fa" exitCode=143 Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.236130 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e850dc8-3f10-4136-b3e6-8b1c13809888","Type":"ContainerDied","Data":"51ef310c1ede4681abb76c58539854417a7ed1b944e97e43716b31a8d0482342"} Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.236153 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e850dc8-3f10-4136-b3e6-8b1c13809888","Type":"ContainerDied","Data":"3d959c8a63aa449f5dc6a40de75353068c1b01e47255024fb3e4b96e593aa1fa"} Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.238047 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerName="glance-log" containerID="cri-o://6a8d2051dd19e9a1d1e84a983702d898b445a62dbb102288d6ee49faf5680a83" gracePeriod=30 Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.238147 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerName="glance-httpd" containerID="cri-o://805a1a6d8892a22250a6b724c2f449cfa72a22d95b42d455070dbad3f696acbf" gracePeriod=30 Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.237921 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9341083c-6983-497b-a0a7-fcca75c3ba65","Type":"ContainerStarted","Data":"805a1a6d8892a22250a6b724c2f449cfa72a22d95b42d455070dbad3f696acbf"} Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.250078 4969 generic.go:334] "Generic (PLEG): container finished" podID="35c12eab-f598-4088-b689-ad192cd83887" containerID="b81188eb7053c4f9e2356cb249f9e3eefc37024c49e128545314631891193999" exitCode=0 Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.250133 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-rskkf" event={"ID":"35c12eab-f598-4088-b689-ad192cd83887","Type":"ContainerDied","Data":"b81188eb7053c4f9e2356cb249f9e3eefc37024c49e128545314631891193999"} Oct 04 08:34:08 crc kubenswrapper[4969]: I1004 08:34:08.286406 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.28637083 podStartE2EDuration="7.28637083s" podCreationTimestamp="2025-10-04 08:34:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:08.259751757 +0000 UTC m=+1076.014020571" watchObservedRunningTime="2025-10-04 08:34:08.28637083 +0000 UTC m=+1076.040639654" Oct 04 08:34:09 crc kubenswrapper[4969]: I1004 08:34:09.064834 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 04 08:34:09 crc kubenswrapper[4969]: I1004 08:34:09.064941 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 04 08:34:09 crc kubenswrapper[4969]: I1004 08:34:09.070013 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 04 08:34:09 crc kubenswrapper[4969]: I1004 08:34:09.263788 4969 generic.go:334] "Generic (PLEG): container finished" podID="6e8240c5-920e-4cbd-9028-46f47a239f45" containerID="b7e88c823b7a23fb328c159dd65d4e4806e654ea40682085f11dd7ccb3d94b19" exitCode=0 Oct 04 08:34:09 crc kubenswrapper[4969]: I1004 08:34:09.264151 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fkm5h" event={"ID":"6e8240c5-920e-4cbd-9028-46f47a239f45","Type":"ContainerDied","Data":"b7e88c823b7a23fb328c159dd65d4e4806e654ea40682085f11dd7ccb3d94b19"} Oct 04 08:34:09 crc kubenswrapper[4969]: I1004 08:34:09.281254 4969 generic.go:334] "Generic (PLEG): container finished" podID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerID="805a1a6d8892a22250a6b724c2f449cfa72a22d95b42d455070dbad3f696acbf" exitCode=0 Oct 04 08:34:09 crc kubenswrapper[4969]: I1004 08:34:09.281294 4969 generic.go:334] "Generic (PLEG): container finished" podID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerID="6a8d2051dd19e9a1d1e84a983702d898b445a62dbb102288d6ee49faf5680a83" exitCode=143 Oct 04 08:34:09 crc kubenswrapper[4969]: I1004 08:34:09.281574 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9341083c-6983-497b-a0a7-fcca75c3ba65","Type":"ContainerDied","Data":"805a1a6d8892a22250a6b724c2f449cfa72a22d95b42d455070dbad3f696acbf"} Oct 04 08:34:09 crc kubenswrapper[4969]: I1004 08:34:09.281628 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9341083c-6983-497b-a0a7-fcca75c3ba65","Type":"ContainerDied","Data":"6a8d2051dd19e9a1d1e84a983702d898b445a62dbb102288d6ee49faf5680a83"} Oct 04 08:34:12 crc kubenswrapper[4969]: I1004 08:34:12.262685 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:34:12 crc kubenswrapper[4969]: I1004 08:34:12.331322 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cdb94bc5-xjwz9"] Oct 04 08:34:12 crc kubenswrapper[4969]: I1004 08:34:12.331608 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerName="dnsmasq-dns" containerID="cri-o://58defe36623f82a68e5be64593dce1537b2a58a430849b278909c50ee12e9745" gracePeriod=10 Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.324837 4969 generic.go:334] "Generic (PLEG): container finished" podID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerID="58defe36623f82a68e5be64593dce1537b2a58a430849b278909c50ee12e9745" exitCode=0 Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.324872 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" event={"ID":"63362185-ca5e-442d-8684-ee9a819ddb9e","Type":"ContainerDied","Data":"58defe36623f82a68e5be64593dce1537b2a58a430849b278909c50ee12e9745"} Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.788902 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.884933 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79cb9c979c-pr7fn"] Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.923979 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6bfr\" (UniqueName: \"kubernetes.io/projected/6e8240c5-920e-4cbd-9028-46f47a239f45-kube-api-access-r6bfr\") pod \"6e8240c5-920e-4cbd-9028-46f47a239f45\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.924340 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-scripts\") pod \"6e8240c5-920e-4cbd-9028-46f47a239f45\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.924394 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-fernet-keys\") pod \"6e8240c5-920e-4cbd-9028-46f47a239f45\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.924532 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-combined-ca-bundle\") pod \"6e8240c5-920e-4cbd-9028-46f47a239f45\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.924574 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-credential-keys\") pod \"6e8240c5-920e-4cbd-9028-46f47a239f45\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.924619 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-config-data\") pod \"6e8240c5-920e-4cbd-9028-46f47a239f45\" (UID: \"6e8240c5-920e-4cbd-9028-46f47a239f45\") " Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.966652 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e8240c5-920e-4cbd-9028-46f47a239f45-kube-api-access-r6bfr" (OuterVolumeSpecName: "kube-api-access-r6bfr") pod "6e8240c5-920e-4cbd-9028-46f47a239f45" (UID: "6e8240c5-920e-4cbd-9028-46f47a239f45"). InnerVolumeSpecName "kube-api-access-r6bfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.985578 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-scripts" (OuterVolumeSpecName: "scripts") pod "6e8240c5-920e-4cbd-9028-46f47a239f45" (UID: "6e8240c5-920e-4cbd-9028-46f47a239f45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:13 crc kubenswrapper[4969]: I1004 08:34:13.989722 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6e8240c5-920e-4cbd-9028-46f47a239f45" (UID: "6e8240c5-920e-4cbd-9028-46f47a239f45"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:13.995494 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6e8240c5-920e-4cbd-9028-46f47a239f45" (UID: "6e8240c5-920e-4cbd-9028-46f47a239f45"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:13.999530 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-586c975586-mxhpd"] Oct 04 08:34:14 crc kubenswrapper[4969]: E1004 08:34:14.000406 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e8240c5-920e-4cbd-9028-46f47a239f45" containerName="keystone-bootstrap" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.000445 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e8240c5-920e-4cbd-9028-46f47a239f45" containerName="keystone-bootstrap" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.000906 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e8240c5-920e-4cbd-9028-46f47a239f45" containerName="keystone-bootstrap" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.002759 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.009656 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.029875 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6bfr\" (UniqueName: \"kubernetes.io/projected/6e8240c5-920e-4cbd-9028-46f47a239f45-kube-api-access-r6bfr\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.029923 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.029937 4969 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.029950 4969 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.058708 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-586c975586-mxhpd"] Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.066538 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e8240c5-920e-4cbd-9028-46f47a239f45" (UID: "6e8240c5-920e-4cbd-9028-46f47a239f45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.118594 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-config-data" (OuterVolumeSpecName: "config-data") pod "6e8240c5-920e-4cbd-9028-46f47a239f45" (UID: "6e8240c5-920e-4cbd-9028-46f47a239f45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.135724 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c876d6d6f-vshkz"] Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.137261 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-secret-key\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.137291 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-combined-ca-bundle\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.137341 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-scripts\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.137392 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99936cdc-7b05-421f-89e9-6e65ee6e2075-logs\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.137453 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc6zz\" (UniqueName: \"kubernetes.io/projected/99936cdc-7b05-421f-89e9-6e65ee6e2075-kube-api-access-zc6zz\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.137476 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-config-data\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.137507 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-tls-certs\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.137568 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.137580 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8240c5-920e-4cbd-9028-46f47a239f45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.170974 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f94787ff8-mkj4z"] Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.173352 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.204752 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f94787ff8-mkj4z"] Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.240772 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3344b512-4dea-4112-a8b1-7fa337ef81ae-scripts\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.240843 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc6zz\" (UniqueName: \"kubernetes.io/projected/99936cdc-7b05-421f-89e9-6e65ee6e2075-kube-api-access-zc6zz\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.240870 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3344b512-4dea-4112-a8b1-7fa337ef81ae-horizon-secret-key\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.240897 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-config-data\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.240917 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3344b512-4dea-4112-a8b1-7fa337ef81ae-logs\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.240961 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-tls-certs\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.240984 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3344b512-4dea-4112-a8b1-7fa337ef81ae-horizon-tls-certs\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.241040 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-secret-key\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.241066 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-combined-ca-bundle\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.241113 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-scripts\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.241139 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82p72\" (UniqueName: \"kubernetes.io/projected/3344b512-4dea-4112-a8b1-7fa337ef81ae-kube-api-access-82p72\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.241183 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3344b512-4dea-4112-a8b1-7fa337ef81ae-config-data\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.241207 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3344b512-4dea-4112-a8b1-7fa337ef81ae-combined-ca-bundle\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.241237 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99936cdc-7b05-421f-89e9-6e65ee6e2075-logs\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.242891 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99936cdc-7b05-421f-89e9-6e65ee6e2075-logs\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.243086 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-config-data\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.245111 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-scripts\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.245773 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-tls-certs\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.246107 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-secret-key\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.246496 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-combined-ca-bundle\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.262148 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc6zz\" (UniqueName: \"kubernetes.io/projected/99936cdc-7b05-421f-89e9-6e65ee6e2075-kube-api-access-zc6zz\") pod \"horizon-586c975586-mxhpd\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.335091 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fkm5h" event={"ID":"6e8240c5-920e-4cbd-9028-46f47a239f45","Type":"ContainerDied","Data":"6008a5d1a943387685fb479d8c2ef8bf6344c82ccc4309e74bf45684bb2becac"} Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.335128 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6008a5d1a943387685fb479d8c2ef8bf6344c82ccc4309e74bf45684bb2becac" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.335199 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fkm5h" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.342137 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3344b512-4dea-4112-a8b1-7fa337ef81ae-horizon-tls-certs\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.342247 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82p72\" (UniqueName: \"kubernetes.io/projected/3344b512-4dea-4112-a8b1-7fa337ef81ae-kube-api-access-82p72\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.342295 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3344b512-4dea-4112-a8b1-7fa337ef81ae-config-data\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.342393 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3344b512-4dea-4112-a8b1-7fa337ef81ae-combined-ca-bundle\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.342465 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3344b512-4dea-4112-a8b1-7fa337ef81ae-scripts\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.342725 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3344b512-4dea-4112-a8b1-7fa337ef81ae-horizon-secret-key\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.342786 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3344b512-4dea-4112-a8b1-7fa337ef81ae-logs\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.343961 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3344b512-4dea-4112-a8b1-7fa337ef81ae-scripts\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.344110 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3344b512-4dea-4112-a8b1-7fa337ef81ae-logs\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.344343 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3344b512-4dea-4112-a8b1-7fa337ef81ae-config-data\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.345280 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3344b512-4dea-4112-a8b1-7fa337ef81ae-horizon-tls-certs\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.346458 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3344b512-4dea-4112-a8b1-7fa337ef81ae-combined-ca-bundle\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.346707 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3344b512-4dea-4112-a8b1-7fa337ef81ae-horizon-secret-key\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.369162 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82p72\" (UniqueName: \"kubernetes.io/projected/3344b512-4dea-4112-a8b1-7fa337ef81ae-kube-api-access-82p72\") pod \"horizon-7f94787ff8-mkj4z\" (UID: \"3344b512-4dea-4112-a8b1-7fa337ef81ae\") " pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.425765 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.495371 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.949443 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fkm5h"] Oct 04 08:34:14 crc kubenswrapper[4969]: I1004 08:34:14.956859 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fkm5h"] Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.052632 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6bdc9"] Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.053818 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.055859 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.056033 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.060282 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.060345 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-82wwh" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.078638 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e8240c5-920e-4cbd-9028-46f47a239f45" path="/var/lib/kubelet/pods/6e8240c5-920e-4cbd-9028-46f47a239f45/volumes" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.079326 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6bdc9"] Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.257477 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-credential-keys\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.257525 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xttqw\" (UniqueName: \"kubernetes.io/projected/e20b7533-fc8d-41c4-be2d-d117a56fa332-kube-api-access-xttqw\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.257574 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-config-data\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.258321 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-combined-ca-bundle\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.258434 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-scripts\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.258513 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-fernet-keys\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.361665 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-credential-keys\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.361707 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xttqw\" (UniqueName: \"kubernetes.io/projected/e20b7533-fc8d-41c4-be2d-d117a56fa332-kube-api-access-xttqw\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.361743 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-config-data\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.361777 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-combined-ca-bundle\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.361794 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-scripts\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.361818 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-fernet-keys\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.367770 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-config-data\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.368859 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-scripts\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.368943 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-combined-ca-bundle\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.375956 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-credential-keys\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.381156 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-fernet-keys\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.382867 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xttqw\" (UniqueName: \"kubernetes.io/projected/e20b7533-fc8d-41c4-be2d-d117a56fa332-kube-api-access-xttqw\") pod \"keystone-bootstrap-6bdc9\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:15 crc kubenswrapper[4969]: I1004 08:34:15.675081 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:17 crc kubenswrapper[4969]: I1004 08:34:17.070071 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: connect: connection refused" Oct 04 08:34:18 crc kubenswrapper[4969]: E1004 08:34:18.061666 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-placement-api:watcher_latest" Oct 04 08:34:18 crc kubenswrapper[4969]: E1004 08:34:18.061713 4969 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-placement-api:watcher_latest" Oct 04 08:34:18 crc kubenswrapper[4969]: E1004 08:34:18.061826 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:38.102.83.83:5001/podified-master-centos10/openstack-placement-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bfl4h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-nx8dt_openstack(4a60a257-4654-4bfd-b878-94733004ac3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 08:34:18 crc kubenswrapper[4969]: E1004 08:34:18.062990 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-nx8dt" podUID="4a60a257-4654-4bfd-b878-94733004ac3d" Oct 04 08:34:18 crc kubenswrapper[4969]: E1004 08:34:18.372579 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.83:5001/podified-master-centos10/openstack-placement-api:watcher_latest\\\"\"" pod="openstack/placement-db-sync-nx8dt" podUID="4a60a257-4654-4bfd-b878-94733004ac3d" Oct 04 08:34:21 crc kubenswrapper[4969]: E1004 08:34:21.974614 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 04 08:34:21 crc kubenswrapper[4969]: E1004 08:34:21.975078 4969 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 04 08:34:21 crc kubenswrapper[4969]: E1004 08:34:21.975300 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n658h8h56fhc9h6fh688h578h5b4h5f8hcfh8ch6bh5c5h659h694h77h7fh5bdh668h5fbh686h546h7bh9bh556h5cfh685h594h548h586h5d6h56fq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-plcfr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-59494b9bd5-6j6nz_openstack(c95b2ea8-1562-44c2-81ad-e4dcbb59cd32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 08:34:21 crc kubenswrapper[4969]: E1004 08:34:21.977500 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-59494b9bd5-6j6nz" podUID="c95b2ea8-1562-44c2-81ad-e4dcbb59cd32" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.014245 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.043687 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.062854 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.062909 4969 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.063038 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65bh664h66fh548hd5hc4h5d9hfh7ch659h689h567h665h65h5cdh5bh5d5h98h8h8bh646h64ch67ch557h6ch4h5c7h597h87h79h5b9h6dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bpfwq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-79cb9c979c-pr7fn_openstack(bf8ccf73-5c59-4a3f-b586-66a7dbea8c49): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.070598 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: connect: connection refused" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.110693 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-79cb9c979c-pr7fn" podUID="bf8ccf73-5c59-4a3f-b586-66a7dbea8c49" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.120566 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.120909 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-rskkf" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.122079 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.194964 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.195022 4969 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.195148 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64bh675h5bbhfbh688hdh8hddh549h665h564h58fh5fbh674h5c7h559h4h65bhb9h557h84h666h65bhbdhb8h65bh58dhc7h94h664h588h544q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wbj29,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-c876d6d6f-vshkz_openstack(df3fb747-e585-4f1f-b0b5-b7fa3ef30f75): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.213957 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-combined-ca-bundle\") pod \"35c12eab-f598-4088-b689-ad192cd83887\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214087 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g46hx\" (UniqueName: \"kubernetes.io/projected/35c12eab-f598-4088-b689-ad192cd83887-kube-api-access-g46hx\") pod \"35c12eab-f598-4088-b689-ad192cd83887\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214115 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-scripts\") pod \"1e850dc8-3f10-4136-b3e6-8b1c13809888\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214139 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-logs\") pod \"1e850dc8-3f10-4136-b3e6-8b1c13809888\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214165 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-httpd-run\") pod \"1e850dc8-3f10-4136-b3e6-8b1c13809888\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214207 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-config-data\") pod \"1e850dc8-3f10-4136-b3e6-8b1c13809888\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214237 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-db-sync-config-data\") pod \"35c12eab-f598-4088-b689-ad192cd83887\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214293 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-combined-ca-bundle\") pod \"1e850dc8-3f10-4136-b3e6-8b1c13809888\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214336 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"1e850dc8-3f10-4136-b3e6-8b1c13809888\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214361 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kpt2\" (UniqueName: \"kubernetes.io/projected/1e850dc8-3f10-4136-b3e6-8b1c13809888-kube-api-access-5kpt2\") pod \"1e850dc8-3f10-4136-b3e6-8b1c13809888\" (UID: \"1e850dc8-3f10-4136-b3e6-8b1c13809888\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.214443 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-config-data\") pod \"35c12eab-f598-4088-b689-ad192cd83887\" (UID: \"35c12eab-f598-4088-b689-ad192cd83887\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.217180 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1e850dc8-3f10-4136-b3e6-8b1c13809888" (UID: "1e850dc8-3f10-4136-b3e6-8b1c13809888"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.228456 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.83:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-c876d6d6f-vshkz" podUID="df3fb747-e585-4f1f-b0b5-b7fa3ef30f75" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.229262 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-logs" (OuterVolumeSpecName: "logs") pod "1e850dc8-3f10-4136-b3e6-8b1c13809888" (UID: "1e850dc8-3f10-4136-b3e6-8b1c13809888"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.243378 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "1e850dc8-3f10-4136-b3e6-8b1c13809888" (UID: "1e850dc8-3f10-4136-b3e6-8b1c13809888"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.243619 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "35c12eab-f598-4088-b689-ad192cd83887" (UID: "35c12eab-f598-4088-b689-ad192cd83887"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.243722 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-scripts" (OuterVolumeSpecName: "scripts") pod "1e850dc8-3f10-4136-b3e6-8b1c13809888" (UID: "1e850dc8-3f10-4136-b3e6-8b1c13809888"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.243793 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35c12eab-f598-4088-b689-ad192cd83887-kube-api-access-g46hx" (OuterVolumeSpecName: "kube-api-access-g46hx") pod "35c12eab-f598-4088-b689-ad192cd83887" (UID: "35c12eab-f598-4088-b689-ad192cd83887"). InnerVolumeSpecName "kube-api-access-g46hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.251689 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e850dc8-3f10-4136-b3e6-8b1c13809888-kube-api-access-5kpt2" (OuterVolumeSpecName: "kube-api-access-5kpt2") pod "1e850dc8-3f10-4136-b3e6-8b1c13809888" (UID: "1e850dc8-3f10-4136-b3e6-8b1c13809888"). InnerVolumeSpecName "kube-api-access-5kpt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.259177 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e850dc8-3f10-4136-b3e6-8b1c13809888" (UID: "1e850dc8-3f10-4136-b3e6-8b1c13809888"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.261822 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35c12eab-f598-4088-b689-ad192cd83887" (UID: "35c12eab-f598-4088-b689-ad192cd83887"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.316629 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.317593 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g46hx\" (UniqueName: \"kubernetes.io/projected/35c12eab-f598-4088-b689-ad192cd83887-kube-api-access-g46hx\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.317608 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.317616 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.317626 4969 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e850dc8-3f10-4136-b3e6-8b1c13809888-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.317634 4969 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.317644 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.317666 4969 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.317676 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kpt2\" (UniqueName: \"kubernetes.io/projected/1e850dc8-3f10-4136-b3e6-8b1c13809888-kube-api-access-5kpt2\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.362629 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-config-data" (OuterVolumeSpecName: "config-data") pod "1e850dc8-3f10-4136-b3e6-8b1c13809888" (UID: "1e850dc8-3f10-4136-b3e6-8b1c13809888"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.381183 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-config-data" (OuterVolumeSpecName: "config-data") pod "35c12eab-f598-4088-b689-ad192cd83887" (UID: "35c12eab-f598-4088-b689-ad192cd83887"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.381836 4969 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.435525 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e850dc8-3f10-4136-b3e6-8b1c13809888-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.435562 4969 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.435571 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c12eab-f598-4088-b689-ad192cd83887-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.442850 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d9f0-account-create-5zlxv" event={"ID":"93ccd81b-783b-4c03-b564-425601b5df4c","Type":"ContainerStarted","Data":"3318d818133f237dfde1d6b3dd75146088c756a50ff1dc03e709505de87bb25d"} Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.448537 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ab2-account-create-mpxpw" event={"ID":"a873063d-e103-48c3-b85e-1b8b2e0ba249","Type":"ContainerStarted","Data":"903fc544a4983c3b08713e9f241a14daebdc7ae037556c1e91f3f72e9673d83f"} Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.455154 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-rskkf" event={"ID":"35c12eab-f598-4088-b689-ad192cd83887","Type":"ContainerDied","Data":"cad6697dcbc70d060504cd3ce8618cf10fa648144f154fe9f8d89491030f05ed"} Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.455234 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cad6697dcbc70d060504cd3ce8618cf10fa648144f154fe9f8d89491030f05ed" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.455203 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-rskkf" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.462791 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e850dc8-3f10-4136-b3e6-8b1c13809888","Type":"ContainerDied","Data":"43ef672942d27c5294361aacee95297a94289e1d99380c5f68f157c3da7ae115"} Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.462842 4969 scope.go:117] "RemoveContainer" containerID="51ef310c1ede4681abb76c58539854417a7ed1b944e97e43716b31a8d0482342" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.472826 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0dbb-account-create-k45jh" event={"ID":"ef5c63b3-e8f7-4b9a-bb7c-642516049ee5","Type":"ContainerStarted","Data":"75e97a7228fb25f50a4a0eeae68ea313ce7a049741e11f0508a216c03b451436"} Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.473449 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.493116 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.646300 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-swift-storage-0\") pod \"63362185-ca5e-442d-8684-ee9a819ddb9e\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.646399 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-config\") pod \"63362185-ca5e-442d-8684-ee9a819ddb9e\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.646470 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-nb\") pod \"63362185-ca5e-442d-8684-ee9a819ddb9e\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.646492 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zkn9\" (UniqueName: \"kubernetes.io/projected/63362185-ca5e-442d-8684-ee9a819ddb9e-kube-api-access-4zkn9\") pod \"63362185-ca5e-442d-8684-ee9a819ddb9e\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.646590 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-sb\") pod \"63362185-ca5e-442d-8684-ee9a819ddb9e\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.646631 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-svc\") pod \"63362185-ca5e-442d-8684-ee9a819ddb9e\" (UID: \"63362185-ca5e-442d-8684-ee9a819ddb9e\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.657982 4969 scope.go:117] "RemoveContainer" containerID="3d959c8a63aa449f5dc6a40de75353068c1b01e47255024fb3e4b96e593aa1fa" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.696661 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.710668 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.715613 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63362185-ca5e-442d-8684-ee9a819ddb9e-kube-api-access-4zkn9" (OuterVolumeSpecName: "kube-api-access-4zkn9") pod "63362185-ca5e-442d-8684-ee9a819ddb9e" (UID: "63362185-ca5e-442d-8684-ee9a819ddb9e"). InnerVolumeSpecName "kube-api-access-4zkn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.743583 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.747725 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-config-data\") pod \"9341083c-6983-497b-a0a7-fcca75c3ba65\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.747774 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-combined-ca-bundle\") pod \"9341083c-6983-497b-a0a7-fcca75c3ba65\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.747907 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"9341083c-6983-497b-a0a7-fcca75c3ba65\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.747928 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-httpd-run\") pod \"9341083c-6983-497b-a0a7-fcca75c3ba65\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.747946 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-scripts\") pod \"9341083c-6983-497b-a0a7-fcca75c3ba65\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.747965 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-logs\") pod \"9341083c-6983-497b-a0a7-fcca75c3ba65\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.748011 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4vc7\" (UniqueName: \"kubernetes.io/projected/9341083c-6983-497b-a0a7-fcca75c3ba65-kube-api-access-x4vc7\") pod \"9341083c-6983-497b-a0a7-fcca75c3ba65\" (UID: \"9341083c-6983-497b-a0a7-fcca75c3ba65\") " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.748531 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zkn9\" (UniqueName: \"kubernetes.io/projected/63362185-ca5e-442d-8684-ee9a819ddb9e-kube-api-access-4zkn9\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.756754 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9341083c-6983-497b-a0a7-fcca75c3ba65" (UID: "9341083c-6983-497b-a0a7-fcca75c3ba65"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.758525 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-logs" (OuterVolumeSpecName: "logs") pod "9341083c-6983-497b-a0a7-fcca75c3ba65" (UID: "9341083c-6983-497b-a0a7-fcca75c3ba65"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.802326 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.802869 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerName="glance-log" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.802890 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerName="glance-log" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.802907 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerName="init" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.802914 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerName="init" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.802925 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c12eab-f598-4088-b689-ad192cd83887" containerName="watcher-db-sync" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.802931 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c12eab-f598-4088-b689-ad192cd83887" containerName="watcher-db-sync" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.802941 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerName="glance-log" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.802948 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerName="glance-log" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.802969 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerName="glance-httpd" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.802975 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerName="glance-httpd" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.802988 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerName="dnsmasq-dns" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.802993 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerName="dnsmasq-dns" Oct 04 08:34:22 crc kubenswrapper[4969]: E1004 08:34:22.803006 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerName="glance-httpd" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.803011 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerName="glance-httpd" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.803168 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" containerName="dnsmasq-dns" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.803180 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerName="glance-log" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.803192 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerName="glance-httpd" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.803199 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c12eab-f598-4088-b689-ad192cd83887" containerName="watcher-db-sync" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.803208 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e850dc8-3f10-4136-b3e6-8b1c13809888" containerName="glance-log" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.803217 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9341083c-6983-497b-a0a7-fcca75c3ba65" containerName="glance-httpd" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.804211 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.810162 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.810356 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.814696 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "9341083c-6983-497b-a0a7-fcca75c3ba65" (UID: "9341083c-6983-497b-a0a7-fcca75c3ba65"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.815502 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-scripts" (OuterVolumeSpecName: "scripts") pod "9341083c-6983-497b-a0a7-fcca75c3ba65" (UID: "9341083c-6983-497b-a0a7-fcca75c3ba65"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.817975 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9341083c-6983-497b-a0a7-fcca75c3ba65-kube-api-access-x4vc7" (OuterVolumeSpecName: "kube-api-access-x4vc7") pod "9341083c-6983-497b-a0a7-fcca75c3ba65" (UID: "9341083c-6983-497b-a0a7-fcca75c3ba65"). InnerVolumeSpecName "kube-api-access-x4vc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.842481 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.853712 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-logs\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.853765 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.853793 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.853843 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4grdx\" (UniqueName: \"kubernetes.io/projected/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-kube-api-access-4grdx\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.853870 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.853904 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.853919 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.853949 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.854000 4969 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.854012 4969 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.854021 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.854029 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9341083c-6983-497b-a0a7-fcca75c3ba65-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.854037 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4vc7\" (UniqueName: \"kubernetes.io/projected/9341083c-6983-497b-a0a7-fcca75c3ba65-kube-api-access-x4vc7\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.960179 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.960432 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.960471 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.960528 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-logs\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.960550 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.960575 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.960618 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4grdx\" (UniqueName: \"kubernetes.io/projected/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-kube-api-access-4grdx\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.960642 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.961612 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-logs\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.961879 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.962443 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.983979 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.984800 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:22 crc kubenswrapper[4969]: I1004 08:34:22.985264 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.013781 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4grdx\" (UniqueName: \"kubernetes.io/projected/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-kube-api-access-4grdx\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.013925 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.091348 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e850dc8-3f10-4136-b3e6-8b1c13809888" path="/var/lib/kubelet/pods/1e850dc8-3f10-4136-b3e6-8b1c13809888/volumes" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.282960 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-config" (OuterVolumeSpecName: "config") pod "63362185-ca5e-442d-8684-ee9a819ddb9e" (UID: "63362185-ca5e-442d-8684-ee9a819ddb9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.319354 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63362185-ca5e-442d-8684-ee9a819ddb9e" (UID: "63362185-ca5e-442d-8684-ee9a819ddb9e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.353635 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "63362185-ca5e-442d-8684-ee9a819ddb9e" (UID: "63362185-ca5e-442d-8684-ee9a819ddb9e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:23 crc kubenswrapper[4969]: W1004 08:34:23.377637 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3344b512_4dea_4112_a8b1_7fa337ef81ae.slice/crio-68c92ed5a0c1db40b08b14c203d52f3ea2d1dff4d38716866b5c92b7e167f307 WatchSource:0}: Error finding container 68c92ed5a0c1db40b08b14c203d52f3ea2d1dff4d38716866b5c92b7e167f307: Status 404 returned error can't find the container with id 68c92ed5a0c1db40b08b14c203d52f3ea2d1dff4d38716866b5c92b7e167f307 Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.394834 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.394875 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.394885 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.440276 4969 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.497543 4969 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.516755 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.526691 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.533900 4969 generic.go:334] "Generic (PLEG): container finished" podID="ef5c63b3-e8f7-4b9a-bb7c-642516049ee5" containerID="6f07054d682b8c5f0df8bcce9ec08f7801b172020e6186e12ad4b4d2321ccc3c" exitCode=0 Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.589314 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-d9f0-account-create-5zlxv" podStartSLOduration=17.589296714 podStartE2EDuration="17.589296714s" podCreationTimestamp="2025-10-04 08:34:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:23.588900975 +0000 UTC m=+1091.343169789" watchObservedRunningTime="2025-10-04 08:34:23.589296714 +0000 UTC m=+1091.343565528" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.622970 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.628667 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9341083c-6983-497b-a0a7-fcca75c3ba65" (UID: "9341083c-6983-497b-a0a7-fcca75c3ba65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.676618 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "63362185-ca5e-442d-8684-ee9a819ddb9e" (UID: "63362185-ca5e-442d-8684-ee9a819ddb9e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.691141 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-config-data" (OuterVolumeSpecName: "config-data") pod "9341083c-6983-497b-a0a7-fcca75c3ba65" (UID: "9341083c-6983-497b-a0a7-fcca75c3ba65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.695108 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63362185-ca5e-442d-8684-ee9a819ddb9e" (UID: "63362185-ca5e-442d-8684-ee9a819ddb9e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.702732 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.702850 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63362185-ca5e-442d-8684-ee9a819ddb9e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.702926 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.702997 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9341083c-6983-497b-a0a7-fcca75c3ba65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.735168 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.965251 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f94787ff8-mkj4z"] Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.965326 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9341083c-6983-497b-a0a7-fcca75c3ba65","Type":"ContainerDied","Data":"a618fb301b028d4978d5c15802cab6df3fbb10b7618022bde0611f15646cd5ae"} Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.965350 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.966780 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.966802 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0dbb-account-create-k45jh" event={"ID":"ef5c63b3-e8f7-4b9a-bb7c-642516049ee5","Type":"ContainerDied","Data":"6f07054d682b8c5f0df8bcce9ec08f7801b172020e6186e12ad4b4d2321ccc3c"} Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.966818 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d9f0-account-create-5zlxv" event={"ID":"93ccd81b-783b-4c03-b564-425601b5df4c","Type":"ContainerStarted","Data":"906fbe48d89804808f909ec418ecb7797887c8aea9472b1601387ae440795c67"} Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.966829 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f94787ff8-mkj4z" event={"ID":"3344b512-4dea-4112-a8b1-7fa337ef81ae","Type":"ContainerStarted","Data":"68c92ed5a0c1db40b08b14c203d52f3ea2d1dff4d38716866b5c92b7e167f307"} Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.966840 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6bdc9"] Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.966851 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.966886 4969 scope.go:117] "RemoveContainer" containerID="805a1a6d8892a22250a6b724c2f449cfa72a22d95b42d455070dbad3f696acbf" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967101 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967532 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c876d6d6f-vshkz" event={"ID":"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75","Type":"ContainerDied","Data":"fbd9f40bbfa205a8b1d61e8a026306b242faaab6a69b934bfba9ebf1e4be56bf"} Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967550 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbd9f40bbfa205a8b1d61e8a026306b242faaab6a69b934bfba9ebf1e4be56bf" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967561 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967710 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-586c975586-mxhpd"] Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967732 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967750 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" event={"ID":"63362185-ca5e-442d-8684-ee9a819ddb9e","Type":"ContainerDied","Data":"e290e1a74c6f81c717ff9e40b836840f302b294a6fde765d1f611347df0076b0"} Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967764 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cb9c979c-pr7fn" event={"ID":"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49","Type":"ContainerDied","Data":"d948976c7d1b55c4fd8e95ba1b067e17e519e05a0e00d918a56abd4fb5b21c86"} Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967774 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d948976c7d1b55c4fd8e95ba1b067e17e519e05a0e00d918a56abd4fb5b21c86" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.967987 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.969206 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.969353 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-szkr4" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.973890 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.973870 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.974817 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 04 08:34:23 crc kubenswrapper[4969]: I1004 08:34:23.976532 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012075 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd8tz\" (UniqueName: \"kubernetes.io/projected/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-kube-api-access-kd8tz\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012117 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw6gk\" (UniqueName: \"kubernetes.io/projected/7144eb34-59d2-4583-8c46-3e56258966da-kube-api-access-mw6gk\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012153 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012273 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c8z8\" (UniqueName: \"kubernetes.io/projected/149bacf3-1026-4b51-9208-eabcd960e409-kube-api-access-8c8z8\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012292 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-config-data\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012338 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-config-data\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012496 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012570 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/149bacf3-1026-4b51-9208-eabcd960e409-logs\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012593 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7144eb34-59d2-4583-8c46-3e56258966da-logs\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012661 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-config-data\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012704 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-logs\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012793 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012830 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.012857 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.100714 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.114726 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.114800 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/149bacf3-1026-4b51-9208-eabcd960e409-logs\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.114828 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7144eb34-59d2-4583-8c46-3e56258966da-logs\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.114879 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-config-data\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.114919 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-logs\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.114982 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.115009 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.115031 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.115153 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd8tz\" (UniqueName: \"kubernetes.io/projected/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-kube-api-access-kd8tz\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.115184 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw6gk\" (UniqueName: \"kubernetes.io/projected/7144eb34-59d2-4583-8c46-3e56258966da-kube-api-access-mw6gk\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.115224 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.115261 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c8z8\" (UniqueName: \"kubernetes.io/projected/149bacf3-1026-4b51-9208-eabcd960e409-kube-api-access-8c8z8\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.115280 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-config-data\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.115314 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-config-data\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.126114 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/149bacf3-1026-4b51-9208-eabcd960e409-logs\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.126432 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7144eb34-59d2-4583-8c46-3e56258966da-logs\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.128020 4969 scope.go:117] "RemoveContainer" containerID="6a8d2051dd19e9a1d1e84a983702d898b445a62dbb102288d6ee49faf5680a83" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.129947 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-logs\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.168542 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.187819 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.188271 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.204324 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.206464 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.210136 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-config-data\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.222446 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c8z8\" (UniqueName: \"kubernetes.io/projected/149bacf3-1026-4b51-9208-eabcd960e409-kube-api-access-8c8z8\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.222522 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw6gk\" (UniqueName: \"kubernetes.io/projected/7144eb34-59d2-4583-8c46-3e56258966da-kube-api-access-mw6gk\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.222639 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd8tz\" (UniqueName: \"kubernetes.io/projected/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-kube-api-access-kd8tz\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.223001 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-config-data\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.223048 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.223159 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac-config-data\") pod \"watcher-applier-0\" (UID: \"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac\") " pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.223213 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.226676 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.231626 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-logs\") pod \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.231689 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-logs\") pod \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.231727 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-config-data\") pod \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.231762 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpfwq\" (UniqueName: \"kubernetes.io/projected/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-kube-api-access-bpfwq\") pod \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.231808 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbj29\" (UniqueName: \"kubernetes.io/projected/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-kube-api-access-wbj29\") pod \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.231937 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-config-data\") pod \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.231963 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-scripts\") pod \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.232160 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-horizon-secret-key\") pod \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\" (UID: \"df3fb747-e585-4f1f-b0b5-b7fa3ef30f75\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.232201 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-scripts\") pod \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.232230 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-horizon-secret-key\") pod \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\" (UID: \"bf8ccf73-5c59-4a3f-b586-66a7dbea8c49\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.236549 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-scripts" (OuterVolumeSpecName: "scripts") pod "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75" (UID: "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.236556 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-config-data" (OuterVolumeSpecName: "config-data") pod "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49" (UID: "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.236983 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-logs" (OuterVolumeSpecName: "logs") pod "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49" (UID: "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.238997 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-scripts" (OuterVolumeSpecName: "scripts") pod "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49" (UID: "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.239135 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75" (UID: "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.239145 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-logs" (OuterVolumeSpecName: "logs") pod "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75" (UID: "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.239214 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-config-data" (OuterVolumeSpecName: "config-data") pod "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75" (UID: "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.257695 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-kube-api-access-wbj29" (OuterVolumeSpecName: "kube-api-access-wbj29") pod "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75" (UID: "df3fb747-e585-4f1f-b0b5-b7fa3ef30f75"). InnerVolumeSpecName "kube-api-access-wbj29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.260777 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-kube-api-access-bpfwq" (OuterVolumeSpecName: "kube-api-access-bpfwq") pod "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49" (UID: "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49"). InnerVolumeSpecName "kube-api-access-bpfwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.267585 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.304320 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.305852 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.318575 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49" (UID: "bf8ccf73-5c59-4a3f-b586-66a7dbea8c49"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.318920 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.328760 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337047 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-config-data\") pod \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337193 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plcfr\" (UniqueName: \"kubernetes.io/projected/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-kube-api-access-plcfr\") pod \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337244 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-horizon-secret-key\") pod \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337274 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-scripts\") pod \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337331 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-logs\") pod \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\" (UID: \"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32\") " Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337837 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337854 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337865 4969 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337876 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337886 4969 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337895 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337905 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337915 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337925 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpfwq\" (UniqueName: \"kubernetes.io/projected/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49-kube-api-access-bpfwq\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.337936 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbj29\" (UniqueName: \"kubernetes.io/projected/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75-kube-api-access-wbj29\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.339469 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-config-data" (OuterVolumeSpecName: "config-data") pod "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32" (UID: "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.339805 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-scripts" (OuterVolumeSpecName: "scripts") pod "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32" (UID: "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.341661 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-logs" (OuterVolumeSpecName: "logs") pod "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32" (UID: "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.343133 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.343568 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32" (UID: "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.378722 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-kube-api-access-plcfr" (OuterVolumeSpecName: "kube-api-access-plcfr") pod "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32" (UID: "c95b2ea8-1562-44c2-81ad-e4dcbb59cd32"). InnerVolumeSpecName "kube-api-access-plcfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439049 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439093 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-logs\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439119 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttz2n\" (UniqueName: \"kubernetes.io/projected/cebd742d-94a0-4014-8172-7bf50d535215-kube-api-access-ttz2n\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439178 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-config-data\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439208 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439228 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439268 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439300 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-scripts\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439348 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439360 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plcfr\" (UniqueName: \"kubernetes.io/projected/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-kube-api-access-plcfr\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439372 4969 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439381 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.439388 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.501644 4969 scope.go:117] "RemoveContainer" containerID="58defe36623f82a68e5be64593dce1537b2a58a430849b278909c50ee12e9745" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.516885 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.541851 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-config-data\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.541930 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.541969 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.542040 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.542093 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-scripts\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.542145 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.542172 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-logs\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.542205 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttz2n\" (UniqueName: \"kubernetes.io/projected/cebd742d-94a0-4014-8172-7bf50d535215-kube-api-access-ttz2n\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.544292 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.550113 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-logs\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.550894 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.592153 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.592586 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-config-data\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.599994 4969 scope.go:117] "RemoveContainer" containerID="aa7b8a719157409a83ce94cd7a2e1bd6a4d71b6187f5c255388dd918f6447a3f" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.600150 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.602472 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-scripts\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.606356 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.621953 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.622923 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttz2n\" (UniqueName: \"kubernetes.io/projected/cebd742d-94a0-4014-8172-7bf50d535215-kube-api-access-ttz2n\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.635801 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.667770 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " pod="openstack/glance-default-external-api-0" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.674140 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6bdc9" event={"ID":"e20b7533-fc8d-41c4-be2d-d117a56fa332","Type":"ContainerStarted","Data":"ffbeee53b6675b679bc21e1c9b25d325391ddfc1135c84945665a51944daa2ad"} Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.684653 4969 generic.go:334] "Generic (PLEG): container finished" podID="a873063d-e103-48c3-b85e-1b8b2e0ba249" containerID="bb8c5d8a95ca5c483cdb480b53f11cf6594af1c1a88be3da46bd52fc9832ea5f" exitCode=0 Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.685015 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ab2-account-create-mpxpw" event={"ID":"a873063d-e103-48c3-b85e-1b8b2e0ba249","Type":"ContainerDied","Data":"bb8c5d8a95ca5c483cdb480b53f11cf6594af1c1a88be3da46bd52fc9832ea5f"} Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.690743 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62","Type":"ContainerStarted","Data":"78279db8344044668848e865e24b5d5b6c610ca8fdfbfa28b6d35a32aed94673"} Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.702631 4969 generic.go:334] "Generic (PLEG): container finished" podID="93ccd81b-783b-4c03-b564-425601b5df4c" containerID="906fbe48d89804808f909ec418ecb7797887c8aea9472b1601387ae440795c67" exitCode=0 Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.702694 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d9f0-account-create-5zlxv" event={"ID":"93ccd81b-783b-4c03-b564-425601b5df4c","Type":"ContainerDied","Data":"906fbe48d89804808f909ec418ecb7797887c8aea9472b1601387ae440795c67"} Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.706780 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586c975586-mxhpd" event={"ID":"99936cdc-7b05-421f-89e9-6e65ee6e2075","Type":"ContainerStarted","Data":"1ea51f099c2095e8c4ad9a8a37772af38d046e5af3f1dd016991426b6c0807db"} Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.706829 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586c975586-mxhpd" event={"ID":"99936cdc-7b05-421f-89e9-6e65ee6e2075","Type":"ContainerStarted","Data":"ad7176901b69399f44e9724125d3cd89d15fcd51666ca8eb8b47c46edfec6152"} Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.711962 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff1b1453-89a7-4b5e-a55d-246e10fa18c2","Type":"ContainerStarted","Data":"667efa089b3b8da7f77a98d0eec9a4f3d66b2d8f64cbb587b1e99170da6b37ab"} Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.723373 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c876d6d6f-vshkz" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.724525 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59494b9bd5-6j6nz" event={"ID":"c95b2ea8-1562-44c2-81ad-e4dcbb59cd32","Type":"ContainerDied","Data":"c3a6b6a3c39f0040245b522d336b6e0e9993fc1761f15e1c0510ddd8f9044bb6"} Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.724583 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59494b9bd5-6j6nz" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.725012 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cb9c979c-pr7fn" Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.822065 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79cb9c979c-pr7fn"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.831722 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79cb9c979c-pr7fn"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.865472 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59494b9bd5-6j6nz"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.899551 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-59494b9bd5-6j6nz"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.923384 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c876d6d6f-vshkz"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.928129 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c876d6d6f-vshkz"] Oct 04 08:34:24 crc kubenswrapper[4969]: I1004 08:34:24.949217 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.075740 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9341083c-6983-497b-a0a7-fcca75c3ba65" path="/var/lib/kubelet/pods/9341083c-6983-497b-a0a7-fcca75c3ba65/volumes" Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.076517 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf8ccf73-5c59-4a3f-b586-66a7dbea8c49" path="/var/lib/kubelet/pods/bf8ccf73-5c59-4a3f-b586-66a7dbea8c49/volumes" Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.077111 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c95b2ea8-1562-44c2-81ad-e4dcbb59cd32" path="/var/lib/kubelet/pods/c95b2ea8-1562-44c2-81ad-e4dcbb59cd32/volumes" Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.077956 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df3fb747-e585-4f1f-b0b5-b7fa3ef30f75" path="/var/lib/kubelet/pods/df3fb747-e585-4f1f-b0b5-b7fa3ef30f75/volumes" Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.324251 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.337565 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.524322 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 04 08:34:25 crc kubenswrapper[4969]: W1004 08:34:25.666395 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7a2121b_d2c2_4735_adb7_9b2bfb7a1fac.slice/crio-4bcec57547327678a4e815346f76eb9037f5c47de15499a08a0d88df4c2570c3 WatchSource:0}: Error finding container 4bcec57547327678a4e815346f76eb9037f5c47de15499a08a0d88df4c2570c3: Status 404 returned error can't find the container with id 4bcec57547327678a4e815346f76eb9037f5c47de15499a08a0d88df4c2570c3 Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.741911 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"149bacf3-1026-4b51-9208-eabcd960e409","Type":"ContainerStarted","Data":"0d227a9af49aec865f6a859d56014da0fbbe5f3c23139b000eb57b2b0aad18ee"} Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.743841 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62","Type":"ContainerStarted","Data":"aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0"} Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.753927 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac","Type":"ContainerStarted","Data":"4bcec57547327678a4e815346f76eb9037f5c47de15499a08a0d88df4c2570c3"} Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.756169 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0dbb-account-create-k45jh" event={"ID":"ef5c63b3-e8f7-4b9a-bb7c-642516049ee5","Type":"ContainerDied","Data":"75e97a7228fb25f50a4a0eeae68ea313ce7a049741e11f0508a216c03b451436"} Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.756205 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75e97a7228fb25f50a4a0eeae68ea313ce7a049741e11f0508a216c03b451436" Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.758370 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6bdc9" event={"ID":"e20b7533-fc8d-41c4-be2d-d117a56fa332","Type":"ContainerStarted","Data":"5010c1e44d7549010cfe78af5a5c7d1d37af909ead2271f81751d7689e42402b"} Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.761552 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerStarted","Data":"fdd644be47a44d016c267448c3c9ead683598afd4ce87d199bde465d7e7a6b88"} Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.770846 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f94787ff8-mkj4z" event={"ID":"3344b512-4dea-4112-a8b1-7fa337ef81ae","Type":"ContainerStarted","Data":"29a8b469b2ace2f5ae045e3f0f5c68b930e648523b8e1d80f081e29fc0a462bd"} Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.780738 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6bdc9" podStartSLOduration=10.780717188 podStartE2EDuration="10.780717188s" podCreationTimestamp="2025-10-04 08:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:25.776159124 +0000 UTC m=+1093.530427938" watchObservedRunningTime="2025-10-04 08:34:25.780717188 +0000 UTC m=+1093.534986012" Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.850571 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0dbb-account-create-k45jh" Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.882135 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lhh4\" (UniqueName: \"kubernetes.io/projected/ef5c63b3-e8f7-4b9a-bb7c-642516049ee5-kube-api-access-2lhh4\") pod \"ef5c63b3-e8f7-4b9a-bb7c-642516049ee5\" (UID: \"ef5c63b3-e8f7-4b9a-bb7c-642516049ee5\") " Oct 04 08:34:25 crc kubenswrapper[4969]: I1004 08:34:25.890156 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef5c63b3-e8f7-4b9a-bb7c-642516049ee5-kube-api-access-2lhh4" (OuterVolumeSpecName: "kube-api-access-2lhh4") pod "ef5c63b3-e8f7-4b9a-bb7c-642516049ee5" (UID: "ef5c63b3-e8f7-4b9a-bb7c-642516049ee5"). InnerVolumeSpecName "kube-api-access-2lhh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:25.995716 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lhh4\" (UniqueName: \"kubernetes.io/projected/ef5c63b3-e8f7-4b9a-bb7c-642516049ee5-kube-api-access-2lhh4\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.337909 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d9f0-account-create-5zlxv" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.338722 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ab2-account-create-mpxpw" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.409258 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lkjh\" (UniqueName: \"kubernetes.io/projected/93ccd81b-783b-4c03-b564-425601b5df4c-kube-api-access-7lkjh\") pod \"93ccd81b-783b-4c03-b564-425601b5df4c\" (UID: \"93ccd81b-783b-4c03-b564-425601b5df4c\") " Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.410391 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxcbc\" (UniqueName: \"kubernetes.io/projected/a873063d-e103-48c3-b85e-1b8b2e0ba249-kube-api-access-gxcbc\") pod \"a873063d-e103-48c3-b85e-1b8b2e0ba249\" (UID: \"a873063d-e103-48c3-b85e-1b8b2e0ba249\") " Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.414735 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93ccd81b-783b-4c03-b564-425601b5df4c-kube-api-access-7lkjh" (OuterVolumeSpecName: "kube-api-access-7lkjh") pod "93ccd81b-783b-4c03-b564-425601b5df4c" (UID: "93ccd81b-783b-4c03-b564-425601b5df4c"). InnerVolumeSpecName "kube-api-access-7lkjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.426752 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a873063d-e103-48c3-b85e-1b8b2e0ba249-kube-api-access-gxcbc" (OuterVolumeSpecName: "kube-api-access-gxcbc") pod "a873063d-e103-48c3-b85e-1b8b2e0ba249" (UID: "a873063d-e103-48c3-b85e-1b8b2e0ba249"). InnerVolumeSpecName "kube-api-access-gxcbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.512920 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lkjh\" (UniqueName: \"kubernetes.io/projected/93ccd81b-783b-4c03-b564-425601b5df4c-kube-api-access-7lkjh\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.512941 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxcbc\" (UniqueName: \"kubernetes.io/projected/a873063d-e103-48c3-b85e-1b8b2e0ba249-kube-api-access-gxcbc\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.518157 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.806656 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cebd742d-94a0-4014-8172-7bf50d535215","Type":"ContainerStarted","Data":"8111f2dd2b1bd14f9c2f06beca1ae39c0dd79be9cd7294e43a9324b2329f0875"} Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.814677 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d9f0-account-create-5zlxv" event={"ID":"93ccd81b-783b-4c03-b564-425601b5df4c","Type":"ContainerDied","Data":"3318d818133f237dfde1d6b3dd75146088c756a50ff1dc03e709505de87bb25d"} Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.814701 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3318d818133f237dfde1d6b3dd75146088c756a50ff1dc03e709505de87bb25d" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.814754 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d9f0-account-create-5zlxv" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.817855 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586c975586-mxhpd" event={"ID":"99936cdc-7b05-421f-89e9-6e65ee6e2075","Type":"ContainerStarted","Data":"d7a1460398fc911e418d1988d141295bff2e22e358f4021fa1bf97a2960e568e"} Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.847206 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-586c975586-mxhpd" podStartSLOduration=13.847180176 podStartE2EDuration="13.847180176s" podCreationTimestamp="2025-10-04 08:34:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:26.842669824 +0000 UTC m=+1094.596938628" watchObservedRunningTime="2025-10-04 08:34:26.847180176 +0000 UTC m=+1094.601449000" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.857232 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ab2-account-create-mpxpw" event={"ID":"a873063d-e103-48c3-b85e-1b8b2e0ba249","Type":"ContainerDied","Data":"903fc544a4983c3b08713e9f241a14daebdc7ae037556c1e91f3f72e9673d83f"} Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.857725 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="903fc544a4983c3b08713e9f241a14daebdc7ae037556c1e91f3f72e9673d83f" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.857247 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ab2-account-create-mpxpw" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.872689 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f94787ff8-mkj4z" event={"ID":"3344b512-4dea-4112-a8b1-7fa337ef81ae","Type":"ContainerStarted","Data":"85103c436e3f5ef3398ed10742e7a1079f3347f40f92520756f036aac5bd2f41"} Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.923279 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f94787ff8-mkj4z" podStartSLOduration=12.645063934 podStartE2EDuration="12.92325081s" podCreationTimestamp="2025-10-04 08:34:14 +0000 UTC" firstStartedPulling="2025-10-04 08:34:23.400643917 +0000 UTC m=+1091.154912731" lastFinishedPulling="2025-10-04 08:34:23.678830793 +0000 UTC m=+1091.433099607" observedRunningTime="2025-10-04 08:34:26.893875469 +0000 UTC m=+1094.648144283" watchObservedRunningTime="2025-10-04 08:34:26.92325081 +0000 UTC m=+1094.677519634" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.926179 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff1b1453-89a7-4b5e-a55d-246e10fa18c2","Type":"ContainerStarted","Data":"779b827573841530d0d9c1ad5778e8eb0addca0b6cf2069e39c769eb1953d8db"} Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.965741 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"149bacf3-1026-4b51-9208-eabcd960e409","Type":"ContainerStarted","Data":"3c69c1d4abd7a0fd6705c86e1214849291b2d5fe7e6e6a1d150d283f6dbfdc2a"} Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.965793 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"149bacf3-1026-4b51-9208-eabcd960e409","Type":"ContainerStarted","Data":"b6d04d5b0f1a6714bbba1655346038197530bc75656bf81a4491925c3711b8ea"} Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.966749 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.998001 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62","Type":"ContainerStarted","Data":"346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6"} Oct 04 08:34:26 crc kubenswrapper[4969]: I1004 08:34:26.998253 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0dbb-account-create-k45jh" Oct 04 08:34:27 crc kubenswrapper[4969]: I1004 08:34:27.012285 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=4.012260036 podStartE2EDuration="4.012260036s" podCreationTimestamp="2025-10-04 08:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:26.993975191 +0000 UTC m=+1094.748244015" watchObservedRunningTime="2025-10-04 08:34:27.012260036 +0000 UTC m=+1094.766528860" Oct 04 08:34:27 crc kubenswrapper[4969]: I1004 08:34:27.021684 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.02166442 podStartE2EDuration="5.02166442s" podCreationTimestamp="2025-10-04 08:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:27.01724625 +0000 UTC m=+1094.771515064" watchObservedRunningTime="2025-10-04 08:34:27.02166442 +0000 UTC m=+1094.775933254" Oct 04 08:34:28 crc kubenswrapper[4969]: I1004 08:34:28.011663 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cebd742d-94a0-4014-8172-7bf50d535215","Type":"ContainerStarted","Data":"faaf50efeefd4b36dea9c5e760a740a8eb39d398df5fc54ee80173f48c59d07f"} Oct 04 08:34:29 crc kubenswrapper[4969]: I1004 08:34:29.031344 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:34:29 crc kubenswrapper[4969]: I1004 08:34:29.312981 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 04 08:34:29 crc kubenswrapper[4969]: I1004 08:34:29.607243 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 04 08:34:30 crc kubenswrapper[4969]: I1004 08:34:30.056786 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerStarted","Data":"02413601cc856344c0caa6d28e7bc8fd58deaef6164283813479dec8f26b23df"} Oct 04 08:34:30 crc kubenswrapper[4969]: I1004 08:34:30.062328 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac","Type":"ContainerStarted","Data":"40edd74f9c13b26ecbaa0b49961bdd59c96da7f255bc4c510a9cc0b83a41bd15"} Oct 04 08:34:30 crc kubenswrapper[4969]: I1004 08:34:30.072484 4969 generic.go:334] "Generic (PLEG): container finished" podID="e20b7533-fc8d-41c4-be2d-d117a56fa332" containerID="5010c1e44d7549010cfe78af5a5c7d1d37af909ead2271f81751d7689e42402b" exitCode=0 Oct 04 08:34:30 crc kubenswrapper[4969]: I1004 08:34:30.072553 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6bdc9" event={"ID":"e20b7533-fc8d-41c4-be2d-d117a56fa332","Type":"ContainerDied","Data":"5010c1e44d7549010cfe78af5a5c7d1d37af909ead2271f81751d7689e42402b"} Oct 04 08:34:30 crc kubenswrapper[4969]: I1004 08:34:30.080300 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cebd742d-94a0-4014-8172-7bf50d535215","Type":"ContainerStarted","Data":"7ac0f9aad61619cccff2e23bf891c80760c98c28eec34aa9f03324022d0bbb28"} Oct 04 08:34:30 crc kubenswrapper[4969]: I1004 08:34:30.091436 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=3.889880408 podStartE2EDuration="7.090819683s" podCreationTimestamp="2025-10-04 08:34:23 +0000 UTC" firstStartedPulling="2025-10-04 08:34:25.688628565 +0000 UTC m=+1093.442897379" lastFinishedPulling="2025-10-04 08:34:28.88956783 +0000 UTC m=+1096.643836654" observedRunningTime="2025-10-04 08:34:30.080756233 +0000 UTC m=+1097.835025047" watchObservedRunningTime="2025-10-04 08:34:30.090819683 +0000 UTC m=+1097.845088497" Oct 04 08:34:30 crc kubenswrapper[4969]: I1004 08:34:30.119543 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=3.923501926 podStartE2EDuration="7.119526219s" podCreationTimestamp="2025-10-04 08:34:23 +0000 UTC" firstStartedPulling="2025-10-04 08:34:25.688673996 +0000 UTC m=+1093.442942840" lastFinishedPulling="2025-10-04 08:34:28.884698279 +0000 UTC m=+1096.638967133" observedRunningTime="2025-10-04 08:34:30.116047633 +0000 UTC m=+1097.870316447" watchObservedRunningTime="2025-10-04 08:34:30.119526219 +0000 UTC m=+1097.873795033" Oct 04 08:34:30 crc kubenswrapper[4969]: I1004 08:34:30.158321 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.158302604 podStartE2EDuration="6.158302604s" podCreationTimestamp="2025-10-04 08:34:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:30.149118555 +0000 UTC m=+1097.903387379" watchObservedRunningTime="2025-10-04 08:34:30.158302604 +0000 UTC m=+1097.912571418" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.622606 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ktnkt"] Oct 04 08:34:31 crc kubenswrapper[4969]: E1004 08:34:31.623351 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a873063d-e103-48c3-b85e-1b8b2e0ba249" containerName="mariadb-account-create" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.623362 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a873063d-e103-48c3-b85e-1b8b2e0ba249" containerName="mariadb-account-create" Oct 04 08:34:31 crc kubenswrapper[4969]: E1004 08:34:31.623377 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ccd81b-783b-4c03-b564-425601b5df4c" containerName="mariadb-account-create" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.623383 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ccd81b-783b-4c03-b564-425601b5df4c" containerName="mariadb-account-create" Oct 04 08:34:31 crc kubenswrapper[4969]: E1004 08:34:31.623411 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef5c63b3-e8f7-4b9a-bb7c-642516049ee5" containerName="mariadb-account-create" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.627457 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef5c63b3-e8f7-4b9a-bb7c-642516049ee5" containerName="mariadb-account-create" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.627804 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef5c63b3-e8f7-4b9a-bb7c-642516049ee5" containerName="mariadb-account-create" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.627822 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ccd81b-783b-4c03-b564-425601b5df4c" containerName="mariadb-account-create" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.627841 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="a873063d-e103-48c3-b85e-1b8b2e0ba249" containerName="mariadb-account-create" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.628442 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.631347 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8dgl8" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.635885 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.640486 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ktnkt"] Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.733974 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-db-sync-config-data\") pod \"barbican-db-sync-ktnkt\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.734280 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-combined-ca-bundle\") pod \"barbican-db-sync-ktnkt\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.734509 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcdtk\" (UniqueName: \"kubernetes.io/projected/b1bc4bef-f750-4b65-96c5-af742379cf87-kube-api-access-wcdtk\") pod \"barbican-db-sync-ktnkt\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.837096 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-db-sync-config-data\") pod \"barbican-db-sync-ktnkt\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.837171 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-combined-ca-bundle\") pod \"barbican-db-sync-ktnkt\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.837217 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcdtk\" (UniqueName: \"kubernetes.io/projected/b1bc4bef-f750-4b65-96c5-af742379cf87-kube-api-access-wcdtk\") pod \"barbican-db-sync-ktnkt\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.851378 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-db-sync-config-data\") pod \"barbican-db-sync-ktnkt\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.851530 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-combined-ca-bundle\") pod \"barbican-db-sync-ktnkt\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.854076 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcdtk\" (UniqueName: \"kubernetes.io/projected/b1bc4bef-f750-4b65-96c5-af742379cf87-kube-api-access-wcdtk\") pod \"barbican-db-sync-ktnkt\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.904568 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-xjc7p"] Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.906106 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.907459 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xw5dg" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.908133 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.908696 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.917769 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xjc7p"] Oct 04 08:34:31 crc kubenswrapper[4969]: I1004 08:34:31.945723 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.004627 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-cf9ls"] Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.006014 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.010379 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.011050 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.011061 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ct9bq" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.014110 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-cf9ls"] Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.040258 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-db-sync-config-data\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.040292 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-combined-ca-bundle\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.040345 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-scripts\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.040413 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-config-data\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.040473 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st8n9\" (UniqueName: \"kubernetes.io/projected/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-kube-api-access-st8n9\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.040499 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-etc-machine-id\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.142771 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-combined-ca-bundle\") pod \"neutron-db-sync-cf9ls\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.142914 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-config-data\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.142988 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st8n9\" (UniqueName: \"kubernetes.io/projected/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-kube-api-access-st8n9\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.143023 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-config\") pod \"neutron-db-sync-cf9ls\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.143054 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-etc-machine-id\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.143100 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-db-sync-config-data\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.143120 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-combined-ca-bundle\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.143175 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjfvn\" (UniqueName: \"kubernetes.io/projected/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-kube-api-access-bjfvn\") pod \"neutron-db-sync-cf9ls\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.143218 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-scripts\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.144738 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-etc-machine-id\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.168133 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-db-sync-config-data\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.169583 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-scripts\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.169796 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-combined-ca-bundle\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.170781 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-config-data\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.173403 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st8n9\" (UniqueName: \"kubernetes.io/projected/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-kube-api-access-st8n9\") pod \"cinder-db-sync-xjc7p\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.245199 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-combined-ca-bundle\") pod \"neutron-db-sync-cf9ls\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.245684 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-config\") pod \"neutron-db-sync-cf9ls\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.245774 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjfvn\" (UniqueName: \"kubernetes.io/projected/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-kube-api-access-bjfvn\") pod \"neutron-db-sync-cf9ls\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.250464 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-combined-ca-bundle\") pod \"neutron-db-sync-cf9ls\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.256628 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-config\") pod \"neutron-db-sync-cf9ls\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.256783 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.262120 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjfvn\" (UniqueName: \"kubernetes.io/projected/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-kube-api-access-bjfvn\") pod \"neutron-db-sync-cf9ls\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:32 crc kubenswrapper[4969]: I1004 08:34:32.333115 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:34:33 crc kubenswrapper[4969]: I1004 08:34:33.121002 4969 generic.go:334] "Generic (PLEG): container finished" podID="7144eb34-59d2-4583-8c46-3e56258966da" containerID="02413601cc856344c0caa6d28e7bc8fd58deaef6164283813479dec8f26b23df" exitCode=1 Oct 04 08:34:33 crc kubenswrapper[4969]: I1004 08:34:33.121057 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerDied","Data":"02413601cc856344c0caa6d28e7bc8fd58deaef6164283813479dec8f26b23df"} Oct 04 08:34:33 crc kubenswrapper[4969]: I1004 08:34:33.121749 4969 scope.go:117] "RemoveContainer" containerID="02413601cc856344c0caa6d28e7bc8fd58deaef6164283813479dec8f26b23df" Oct 04 08:34:33 crc kubenswrapper[4969]: I1004 08:34:33.736346 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:33 crc kubenswrapper[4969]: I1004 08:34:33.737035 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:33 crc kubenswrapper[4969]: I1004 08:34:33.778268 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:33 crc kubenswrapper[4969]: I1004 08:34:33.793505 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.131694 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.131727 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.427940 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.428402 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.500793 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.501213 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.554699 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.607919 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.620604 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.622977 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.622997 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.637586 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.638436 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.670711 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.705857 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xttqw\" (UniqueName: \"kubernetes.io/projected/e20b7533-fc8d-41c4-be2d-d117a56fa332-kube-api-access-xttqw\") pod \"e20b7533-fc8d-41c4-be2d-d117a56fa332\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.705992 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-scripts\") pod \"e20b7533-fc8d-41c4-be2d-d117a56fa332\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.706097 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-fernet-keys\") pod \"e20b7533-fc8d-41c4-be2d-d117a56fa332\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.706150 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-credential-keys\") pod \"e20b7533-fc8d-41c4-be2d-d117a56fa332\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.706183 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-combined-ca-bundle\") pod \"e20b7533-fc8d-41c4-be2d-d117a56fa332\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.706260 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-config-data\") pod \"e20b7533-fc8d-41c4-be2d-d117a56fa332\" (UID: \"e20b7533-fc8d-41c4-be2d-d117a56fa332\") " Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.711002 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-scripts" (OuterVolumeSpecName: "scripts") pod "e20b7533-fc8d-41c4-be2d-d117a56fa332" (UID: "e20b7533-fc8d-41c4-be2d-d117a56fa332"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.711414 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e20b7533-fc8d-41c4-be2d-d117a56fa332-kube-api-access-xttqw" (OuterVolumeSpecName: "kube-api-access-xttqw") pod "e20b7533-fc8d-41c4-be2d-d117a56fa332" (UID: "e20b7533-fc8d-41c4-be2d-d117a56fa332"). InnerVolumeSpecName "kube-api-access-xttqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.712017 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e20b7533-fc8d-41c4-be2d-d117a56fa332" (UID: "e20b7533-fc8d-41c4-be2d-d117a56fa332"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.712428 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e20b7533-fc8d-41c4-be2d-d117a56fa332" (UID: "e20b7533-fc8d-41c4-be2d-d117a56fa332"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.731693 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-config-data" (OuterVolumeSpecName: "config-data") pod "e20b7533-fc8d-41c4-be2d-d117a56fa332" (UID: "e20b7533-fc8d-41c4-be2d-d117a56fa332"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.736441 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e20b7533-fc8d-41c4-be2d-d117a56fa332" (UID: "e20b7533-fc8d-41c4-be2d-d117a56fa332"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.812178 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.812228 4969 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.812239 4969 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.812250 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.812260 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20b7533-fc8d-41c4-be2d-d117a56fa332-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.812268 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xttqw\" (UniqueName: \"kubernetes.io/projected/e20b7533-fc8d-41c4-be2d-d117a56fa332-kube-api-access-xttqw\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.819477 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ktnkt"] Oct 04 08:34:34 crc kubenswrapper[4969]: W1004 08:34:34.945093 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod278ebf83_6c66_4d97_8b5c_a7a83ace00b7.slice/crio-16bf5704bca58400ccfe6aedaf4914de3e1bf935a3e155ade19202e72333893a WatchSource:0}: Error finding container 16bf5704bca58400ccfe6aedaf4914de3e1bf935a3e155ade19202e72333893a: Status 404 returned error can't find the container with id 16bf5704bca58400ccfe6aedaf4914de3e1bf935a3e155ade19202e72333893a Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.947695 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xjc7p"] Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.950026 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.951481 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.960410 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-cf9ls"] Oct 04 08:34:34 crc kubenswrapper[4969]: I1004 08:34:34.995933 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.034684 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.143845 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ktnkt" event={"ID":"b1bc4bef-f750-4b65-96c5-af742379cf87","Type":"ContainerStarted","Data":"098dc0c3b31cc252000abccd87dff62b330a236a61ef2f28d1c5463c342022b4"} Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.146054 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerStarted","Data":"9c1a4f5d79ad64001bd9157a55394ac39bd51590bc1287dc6daa3348dab0dcd9"} Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.150809 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nx8dt" event={"ID":"4a60a257-4654-4bfd-b878-94733004ac3d","Type":"ContainerStarted","Data":"f22e41ad155ab7a114bc1d62a5a37050455b6bf0e9177b4657828cec48d2a144"} Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.154624 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff1b1453-89a7-4b5e-a55d-246e10fa18c2","Type":"ContainerStarted","Data":"b080b21623eac49fc2e1208866892b6b46784834630fbd4864315ed3db91409a"} Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.156546 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cf9ls" event={"ID":"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf","Type":"ContainerStarted","Data":"fa7d3b44e484120a5a3ea50089d66bad7288ad199d49aca435555dc55b054f41"} Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.156594 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cf9ls" event={"ID":"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf","Type":"ContainerStarted","Data":"13ac15667bf398c16f59ba748239280cf910691d69de0ed27a6b8bbe5073bae2"} Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.157768 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xjc7p" event={"ID":"278ebf83-6c66-4d97-8b5c-a7a83ace00b7","Type":"ContainerStarted","Data":"16bf5704bca58400ccfe6aedaf4914de3e1bf935a3e155ade19202e72333893a"} Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.161132 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6bdc9" event={"ID":"e20b7533-fc8d-41c4-be2d-d117a56fa332","Type":"ContainerDied","Data":"ffbeee53b6675b679bc21e1c9b25d325391ddfc1135c84945665a51944daa2ad"} Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.161206 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffbeee53b6675b679bc21e1c9b25d325391ddfc1135c84945665a51944daa2ad" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.161259 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6bdc9" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.173319 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.173631 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.179680 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.190009 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nx8dt" podStartSLOduration=2.9754457370000003 podStartE2EDuration="34.189993536s" podCreationTimestamp="2025-10-04 08:34:01 +0000 UTC" firstStartedPulling="2025-10-04 08:34:03.106587355 +0000 UTC m=+1070.860856169" lastFinishedPulling="2025-10-04 08:34:34.321135154 +0000 UTC m=+1102.075403968" observedRunningTime="2025-10-04 08:34:35.187107015 +0000 UTC m=+1102.941375849" watchObservedRunningTime="2025-10-04 08:34:35.189993536 +0000 UTC m=+1102.944262350" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.202304 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-cf9ls" podStartSLOduration=4.202280316 podStartE2EDuration="4.202280316s" podCreationTimestamp="2025-10-04 08:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:35.202065641 +0000 UTC m=+1102.956334465" watchObservedRunningTime="2025-10-04 08:34:35.202280316 +0000 UTC m=+1102.956549130" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.214189 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.679068 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-699ff78d6b-dv46q"] Oct 04 08:34:35 crc kubenswrapper[4969]: E1004 08:34:35.679835 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20b7533-fc8d-41c4-be2d-d117a56fa332" containerName="keystone-bootstrap" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.679852 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20b7533-fc8d-41c4-be2d-d117a56fa332" containerName="keystone-bootstrap" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.680088 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="e20b7533-fc8d-41c4-be2d-d117a56fa332" containerName="keystone-bootstrap" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.680824 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.685019 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.686196 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.688927 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-699ff78d6b-dv46q"] Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.689071 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-82wwh" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.689321 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.689443 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.689564 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.739562 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-credential-keys\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.740039 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-fernet-keys\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.740130 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-internal-tls-certs\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.740180 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmmjh\" (UniqueName: \"kubernetes.io/projected/7327643b-aa39-4085-88c4-bc4411b8d832-kube-api-access-bmmjh\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.740220 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-public-tls-certs\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.740271 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-combined-ca-bundle\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.740332 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-config-data\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.740377 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-scripts\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.841738 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-internal-tls-certs\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.841801 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmmjh\" (UniqueName: \"kubernetes.io/projected/7327643b-aa39-4085-88c4-bc4411b8d832-kube-api-access-bmmjh\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.841829 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-public-tls-certs\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.841852 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-combined-ca-bundle\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.841882 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-config-data\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.841909 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-scripts\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.841966 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-credential-keys\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.841999 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-fernet-keys\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.851642 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-scripts\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.855086 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-public-tls-certs\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.856347 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-fernet-keys\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.861284 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-combined-ca-bundle\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.863492 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-internal-tls-certs\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.865284 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmmjh\" (UniqueName: \"kubernetes.io/projected/7327643b-aa39-4085-88c4-bc4411b8d832-kube-api-access-bmmjh\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.866655 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-config-data\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:35 crc kubenswrapper[4969]: I1004 08:34:35.867936 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7327643b-aa39-4085-88c4-bc4411b8d832-credential-keys\") pod \"keystone-699ff78d6b-dv46q\" (UID: \"7327643b-aa39-4085-88c4-bc4411b8d832\") " pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:36 crc kubenswrapper[4969]: I1004 08:34:36.039674 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:36 crc kubenswrapper[4969]: I1004 08:34:36.203621 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:34:36 crc kubenswrapper[4969]: I1004 08:34:36.203638 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:34:36 crc kubenswrapper[4969]: I1004 08:34:36.664064 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-699ff78d6b-dv46q"] Oct 04 08:34:36 crc kubenswrapper[4969]: W1004 08:34:36.681326 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7327643b_aa39_4085_88c4_bc4411b8d832.slice/crio-a6a0a907ab19de823e796ce19839eeea769192459420cd09ab0901213c8df78b WatchSource:0}: Error finding container a6a0a907ab19de823e796ce19839eeea769192459420cd09ab0901213c8df78b: Status 404 returned error can't find the container with id a6a0a907ab19de823e796ce19839eeea769192459420cd09ab0901213c8df78b Oct 04 08:34:37 crc kubenswrapper[4969]: I1004 08:34:37.122986 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:37 crc kubenswrapper[4969]: I1004 08:34:37.123465 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 08:34:37 crc kubenswrapper[4969]: I1004 08:34:37.226472 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-699ff78d6b-dv46q" event={"ID":"7327643b-aa39-4085-88c4-bc4411b8d832","Type":"ContainerStarted","Data":"a0785231870a4b33f838e713b6cdbeadc1bdf3a7608744ef845b0348f00041cd"} Oct 04 08:34:37 crc kubenswrapper[4969]: I1004 08:34:37.226766 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-699ff78d6b-dv46q" event={"ID":"7327643b-aa39-4085-88c4-bc4411b8d832","Type":"ContainerStarted","Data":"a6a0a907ab19de823e796ce19839eeea769192459420cd09ab0901213c8df78b"} Oct 04 08:34:37 crc kubenswrapper[4969]: I1004 08:34:37.226539 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:34:37 crc kubenswrapper[4969]: I1004 08:34:37.226792 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:34:37 crc kubenswrapper[4969]: I1004 08:34:37.227071 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:34:37 crc kubenswrapper[4969]: I1004 08:34:37.248908 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-699ff78d6b-dv46q" podStartSLOduration=2.2488926989999998 podStartE2EDuration="2.248892699s" podCreationTimestamp="2025-10-04 08:34:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:34:37.248060859 +0000 UTC m=+1105.002329673" watchObservedRunningTime="2025-10-04 08:34:37.248892699 +0000 UTC m=+1105.003161513" Oct 04 08:34:37 crc kubenswrapper[4969]: I1004 08:34:37.739572 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 08:34:38 crc kubenswrapper[4969]: I1004 08:34:38.159367 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 04 08:34:38 crc kubenswrapper[4969]: I1004 08:34:38.160069 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api-log" containerID="cri-o://b6d04d5b0f1a6714bbba1655346038197530bc75656bf81a4491925c3711b8ea" gracePeriod=30 Oct 04 08:34:38 crc kubenswrapper[4969]: I1004 08:34:38.160155 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api" containerID="cri-o://3c69c1d4abd7a0fd6705c86e1214849291b2d5fe7e6e6a1d150d283f6dbfdc2a" gracePeriod=30 Oct 04 08:34:38 crc kubenswrapper[4969]: I1004 08:34:38.237726 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:34:38 crc kubenswrapper[4969]: I1004 08:34:38.333313 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 08:34:39 crc kubenswrapper[4969]: I1004 08:34:39.284386 4969 generic.go:334] "Generic (PLEG): container finished" podID="149bacf3-1026-4b51-9208-eabcd960e409" containerID="b6d04d5b0f1a6714bbba1655346038197530bc75656bf81a4491925c3711b8ea" exitCode=143 Oct 04 08:34:39 crc kubenswrapper[4969]: I1004 08:34:39.284481 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"149bacf3-1026-4b51-9208-eabcd960e409","Type":"ContainerDied","Data":"b6d04d5b0f1a6714bbba1655346038197530bc75656bf81a4491925c3711b8ea"} Oct 04 08:34:39 crc kubenswrapper[4969]: E1004 08:34:39.557373 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7144eb34_59d2_4583_8c46_3e56258966da.slice/crio-conmon-9c1a4f5d79ad64001bd9157a55394ac39bd51590bc1287dc6daa3348dab0dcd9.scope\": RecentStats: unable to find data in memory cache]" Oct 04 08:34:39 crc kubenswrapper[4969]: I1004 08:34:39.774116 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.161:9322/\": read tcp 10.217.0.2:40332->10.217.0.161:9322: read: connection reset by peer" Oct 04 08:34:39 crc kubenswrapper[4969]: I1004 08:34:39.774742 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9322/\": read tcp 10.217.0.2:40338->10.217.0.161:9322: read: connection reset by peer" Oct 04 08:34:40 crc kubenswrapper[4969]: I1004 08:34:40.295163 4969 generic.go:334] "Generic (PLEG): container finished" podID="149bacf3-1026-4b51-9208-eabcd960e409" containerID="3c69c1d4abd7a0fd6705c86e1214849291b2d5fe7e6e6a1d150d283f6dbfdc2a" exitCode=0 Oct 04 08:34:40 crc kubenswrapper[4969]: I1004 08:34:40.295223 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"149bacf3-1026-4b51-9208-eabcd960e409","Type":"ContainerDied","Data":"3c69c1d4abd7a0fd6705c86e1214849291b2d5fe7e6e6a1d150d283f6dbfdc2a"} Oct 04 08:34:40 crc kubenswrapper[4969]: I1004 08:34:40.297615 4969 generic.go:334] "Generic (PLEG): container finished" podID="7144eb34-59d2-4583-8c46-3e56258966da" containerID="9c1a4f5d79ad64001bd9157a55394ac39bd51590bc1287dc6daa3348dab0dcd9" exitCode=1 Oct 04 08:34:40 crc kubenswrapper[4969]: I1004 08:34:40.297711 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerDied","Data":"9c1a4f5d79ad64001bd9157a55394ac39bd51590bc1287dc6daa3348dab0dcd9"} Oct 04 08:34:40 crc kubenswrapper[4969]: I1004 08:34:40.297771 4969 scope.go:117] "RemoveContainer" containerID="02413601cc856344c0caa6d28e7bc8fd58deaef6164283813479dec8f26b23df" Oct 04 08:34:40 crc kubenswrapper[4969]: I1004 08:34:40.298357 4969 scope.go:117] "RemoveContainer" containerID="9c1a4f5d79ad64001bd9157a55394ac39bd51590bc1287dc6daa3348dab0dcd9" Oct 04 08:34:40 crc kubenswrapper[4969]: E1004 08:34:40.298588 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:34:40 crc kubenswrapper[4969]: I1004 08:34:40.301086 4969 generic.go:334] "Generic (PLEG): container finished" podID="4a60a257-4654-4bfd-b878-94733004ac3d" containerID="f22e41ad155ab7a114bc1d62a5a37050455b6bf0e9177b4657828cec48d2a144" exitCode=0 Oct 04 08:34:40 crc kubenswrapper[4969]: I1004 08:34:40.301115 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nx8dt" event={"ID":"4a60a257-4654-4bfd-b878-94733004ac3d","Type":"ContainerDied","Data":"f22e41ad155ab7a114bc1d62a5a37050455b6bf0e9177b4657828cec48d2a144"} Oct 04 08:34:44 crc kubenswrapper[4969]: I1004 08:34:44.622310 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:34:44 crc kubenswrapper[4969]: I1004 08:34:44.622893 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:34:44 crc kubenswrapper[4969]: I1004 08:34:44.623328 4969 scope.go:117] "RemoveContainer" containerID="9c1a4f5d79ad64001bd9157a55394ac39bd51590bc1287dc6daa3348dab0dcd9" Oct 04 08:34:44 crc kubenswrapper[4969]: E1004 08:34:44.623589 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:34:46 crc kubenswrapper[4969]: I1004 08:34:46.351290 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:46 crc kubenswrapper[4969]: I1004 08:34:46.448581 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:48 crc kubenswrapper[4969]: I1004 08:34:48.085322 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:34:48 crc kubenswrapper[4969]: I1004 08:34:48.130991 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7f94787ff8-mkj4z" Oct 04 08:34:48 crc kubenswrapper[4969]: I1004 08:34:48.216718 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-586c975586-mxhpd"] Oct 04 08:34:48 crc kubenswrapper[4969]: I1004 08:34:48.403914 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-586c975586-mxhpd" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon-log" containerID="cri-o://1ea51f099c2095e8c4ad9a8a37772af38d046e5af3f1dd016991426b6c0807db" gracePeriod=30 Oct 04 08:34:48 crc kubenswrapper[4969]: I1004 08:34:48.404058 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-586c975586-mxhpd" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon" containerID="cri-o://d7a1460398fc911e418d1988d141295bff2e22e358f4021fa1bf97a2960e568e" gracePeriod=30 Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.179997 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.196646 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.325562 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c8z8\" (UniqueName: \"kubernetes.io/projected/149bacf3-1026-4b51-9208-eabcd960e409-kube-api-access-8c8z8\") pod \"149bacf3-1026-4b51-9208-eabcd960e409\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.325638 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a60a257-4654-4bfd-b878-94733004ac3d-logs\") pod \"4a60a257-4654-4bfd-b878-94733004ac3d\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.325695 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfl4h\" (UniqueName: \"kubernetes.io/projected/4a60a257-4654-4bfd-b878-94733004ac3d-kube-api-access-bfl4h\") pod \"4a60a257-4654-4bfd-b878-94733004ac3d\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.325729 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/149bacf3-1026-4b51-9208-eabcd960e409-logs\") pod \"149bacf3-1026-4b51-9208-eabcd960e409\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.325839 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-config-data\") pod \"4a60a257-4654-4bfd-b878-94733004ac3d\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.325879 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-combined-ca-bundle\") pod \"149bacf3-1026-4b51-9208-eabcd960e409\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.325974 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-config-data\") pod \"149bacf3-1026-4b51-9208-eabcd960e409\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.326039 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-combined-ca-bundle\") pod \"4a60a257-4654-4bfd-b878-94733004ac3d\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.326077 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-custom-prometheus-ca\") pod \"149bacf3-1026-4b51-9208-eabcd960e409\" (UID: \"149bacf3-1026-4b51-9208-eabcd960e409\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.326119 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-scripts\") pod \"4a60a257-4654-4bfd-b878-94733004ac3d\" (UID: \"4a60a257-4654-4bfd-b878-94733004ac3d\") " Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.328950 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/149bacf3-1026-4b51-9208-eabcd960e409-logs" (OuterVolumeSpecName: "logs") pod "149bacf3-1026-4b51-9208-eabcd960e409" (UID: "149bacf3-1026-4b51-9208-eabcd960e409"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.329062 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a60a257-4654-4bfd-b878-94733004ac3d-logs" (OuterVolumeSpecName: "logs") pod "4a60a257-4654-4bfd-b878-94733004ac3d" (UID: "4a60a257-4654-4bfd-b878-94733004ac3d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.333668 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-scripts" (OuterVolumeSpecName: "scripts") pod "4a60a257-4654-4bfd-b878-94733004ac3d" (UID: "4a60a257-4654-4bfd-b878-94733004ac3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.344997 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/149bacf3-1026-4b51-9208-eabcd960e409-kube-api-access-8c8z8" (OuterVolumeSpecName: "kube-api-access-8c8z8") pod "149bacf3-1026-4b51-9208-eabcd960e409" (UID: "149bacf3-1026-4b51-9208-eabcd960e409"). InnerVolumeSpecName "kube-api-access-8c8z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.352259 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a60a257-4654-4bfd-b878-94733004ac3d-kube-api-access-bfl4h" (OuterVolumeSpecName: "kube-api-access-bfl4h") pod "4a60a257-4654-4bfd-b878-94733004ac3d" (UID: "4a60a257-4654-4bfd-b878-94733004ac3d"). InnerVolumeSpecName "kube-api-access-bfl4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.361032 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a60a257-4654-4bfd-b878-94733004ac3d" (UID: "4a60a257-4654-4bfd-b878-94733004ac3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.368733 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "149bacf3-1026-4b51-9208-eabcd960e409" (UID: "149bacf3-1026-4b51-9208-eabcd960e409"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.378219 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-config-data" (OuterVolumeSpecName: "config-data") pod "4a60a257-4654-4bfd-b878-94733004ac3d" (UID: "4a60a257-4654-4bfd-b878-94733004ac3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.392940 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "149bacf3-1026-4b51-9208-eabcd960e409" (UID: "149bacf3-1026-4b51-9208-eabcd960e409"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.397299 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-config-data" (OuterVolumeSpecName: "config-data") pod "149bacf3-1026-4b51-9208-eabcd960e409" (UID: "149bacf3-1026-4b51-9208-eabcd960e409"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.416707 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nx8dt" event={"ID":"4a60a257-4654-4bfd-b878-94733004ac3d","Type":"ContainerDied","Data":"a22ce1ba338ee94f01c93daa4d6480c09adcc077b78c695a9ec5e1cfeaa1a58a"} Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.416746 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a22ce1ba338ee94f01c93daa4d6480c09adcc077b78c695a9ec5e1cfeaa1a58a" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.416799 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nx8dt" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.422101 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"149bacf3-1026-4b51-9208-eabcd960e409","Type":"ContainerDied","Data":"0d227a9af49aec865f6a859d56014da0fbbe5f3c23139b000eb57b2b0aad18ee"} Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.422191 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430824 4969 generic.go:334] "Generic (PLEG): container finished" podID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerID="d7a1460398fc911e418d1988d141295bff2e22e358f4021fa1bf97a2960e568e" exitCode=0 Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430869 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586c975586-mxhpd" event={"ID":"99936cdc-7b05-421f-89e9-6e65ee6e2075","Type":"ContainerDied","Data":"d7a1460398fc911e418d1988d141295bff2e22e358f4021fa1bf97a2960e568e"} Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430904 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c8z8\" (UniqueName: \"kubernetes.io/projected/149bacf3-1026-4b51-9208-eabcd960e409-kube-api-access-8c8z8\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430926 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a60a257-4654-4bfd-b878-94733004ac3d-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430936 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfl4h\" (UniqueName: \"kubernetes.io/projected/4a60a257-4654-4bfd-b878-94733004ac3d-kube-api-access-bfl4h\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430945 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/149bacf3-1026-4b51-9208-eabcd960e409-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430953 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430963 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430971 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430980 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430988 4969 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/149bacf3-1026-4b51-9208-eabcd960e409-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.430997 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a60a257-4654-4bfd-b878-94733004ac3d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.512436 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.521739 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.532329 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 04 08:34:49 crc kubenswrapper[4969]: E1004 08:34:49.533018 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api-log" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.533037 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api-log" Oct 04 08:34:49 crc kubenswrapper[4969]: E1004 08:34:49.533082 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a60a257-4654-4bfd-b878-94733004ac3d" containerName="placement-db-sync" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.533090 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a60a257-4654-4bfd-b878-94733004ac3d" containerName="placement-db-sync" Oct 04 08:34:49 crc kubenswrapper[4969]: E1004 08:34:49.533105 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.533111 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.533274 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.533297 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api-log" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.533308 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a60a257-4654-4bfd-b878-94733004ac3d" containerName="placement-db-sync" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.534242 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.537453 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.537591 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.538473 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.550817 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.608390 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.161:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.608409 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="149bacf3-1026-4b51-9208-eabcd960e409" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.634056 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2p6v\" (UniqueName: \"kubernetes.io/projected/b36dc326-a420-42e1-a274-9e82641ee148-kube-api-access-z2p6v\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.634164 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b36dc326-a420-42e1-a274-9e82641ee148-logs\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.634193 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-config-data\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.634320 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.634393 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.634432 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-public-tls-certs\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.634503 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.735936 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2p6v\" (UniqueName: \"kubernetes.io/projected/b36dc326-a420-42e1-a274-9e82641ee148-kube-api-access-z2p6v\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.736442 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b36dc326-a420-42e1-a274-9e82641ee148-logs\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.736494 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-config-data\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.736535 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.736580 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.736607 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-public-tls-certs\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.736650 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.737627 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b36dc326-a420-42e1-a274-9e82641ee148-logs\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.740048 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-public-tls-certs\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.740736 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.741561 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.741813 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-config-data\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.748843 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36dc326-a420-42e1-a274-9e82641ee148-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.751025 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2p6v\" (UniqueName: \"kubernetes.io/projected/b36dc326-a420-42e1-a274-9e82641ee148-kube-api-access-z2p6v\") pod \"watcher-api-0\" (UID: \"b36dc326-a420-42e1-a274-9e82641ee148\") " pod="openstack/watcher-api-0" Oct 04 08:34:49 crc kubenswrapper[4969]: I1004 08:34:49.865086 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.311271 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-569b746d9d-r7c45"] Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.312690 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.323842 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.324083 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.326990 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.327104 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ssb6j" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.327167 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.335304 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-569b746d9d-r7c45"] Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.455572 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l98c7\" (UniqueName: \"kubernetes.io/projected/d6425541-fb00-4545-a4df-64aba5d7e5ab-kube-api-access-l98c7\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.455951 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-scripts\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.455976 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-combined-ca-bundle\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.456009 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-config-data\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.456091 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-internal-tls-certs\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.456116 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-public-tls-certs\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.456248 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6425541-fb00-4545-a4df-64aba5d7e5ab-logs\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.557684 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6425541-fb00-4545-a4df-64aba5d7e5ab-logs\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.557722 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l98c7\" (UniqueName: \"kubernetes.io/projected/d6425541-fb00-4545-a4df-64aba5d7e5ab-kube-api-access-l98c7\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.557760 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-scripts\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.557776 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-combined-ca-bundle\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.557798 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-config-data\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.557844 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-internal-tls-certs\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.557861 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-public-tls-certs\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.558141 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6425541-fb00-4545-a4df-64aba5d7e5ab-logs\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.563446 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-scripts\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.563891 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-internal-tls-certs\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.564523 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-combined-ca-bundle\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.565632 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-public-tls-certs\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.578098 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6425541-fb00-4545-a4df-64aba5d7e5ab-config-data\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.578736 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l98c7\" (UniqueName: \"kubernetes.io/projected/d6425541-fb00-4545-a4df-64aba5d7e5ab-kube-api-access-l98c7\") pod \"placement-569b746d9d-r7c45\" (UID: \"d6425541-fb00-4545-a4df-64aba5d7e5ab\") " pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:50 crc kubenswrapper[4969]: I1004 08:34:50.641911 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:34:51 crc kubenswrapper[4969]: I1004 08:34:51.067919 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="149bacf3-1026-4b51-9208-eabcd960e409" path="/var/lib/kubelet/pods/149bacf3-1026-4b51-9208-eabcd960e409/volumes" Oct 04 08:34:53 crc kubenswrapper[4969]: I1004 08:34:53.973917 4969 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod63362185-ca5e-442d-8684-ee9a819ddb9e"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod63362185-ca5e-442d-8684-ee9a819ddb9e] : Timed out while waiting for systemd to remove kubepods-besteffort-pod63362185_ca5e_442d_8684_ee9a819ddb9e.slice" Oct 04 08:34:53 crc kubenswrapper[4969]: E1004 08:34:53.974402 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod63362185-ca5e-442d-8684-ee9a819ddb9e] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod63362185-ca5e-442d-8684-ee9a819ddb9e] : Timed out while waiting for systemd to remove kubepods-besteffort-pod63362185_ca5e_442d_8684_ee9a819ddb9e.slice" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" Oct 04 08:34:54 crc kubenswrapper[4969]: I1004 08:34:54.427952 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-586c975586-mxhpd" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Oct 04 08:34:54 crc kubenswrapper[4969]: I1004 08:34:54.479670 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cdb94bc5-xjwz9" Oct 04 08:34:54 crc kubenswrapper[4969]: I1004 08:34:54.507752 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cdb94bc5-xjwz9"] Oct 04 08:34:54 crc kubenswrapper[4969]: I1004 08:34:54.516013 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cdb94bc5-xjwz9"] Oct 04 08:34:54 crc kubenswrapper[4969]: I1004 08:34:54.626872 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 04 08:34:54 crc kubenswrapper[4969]: I1004 08:34:54.627124 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:34:54 crc kubenswrapper[4969]: I1004 08:34:54.628357 4969 scope.go:117] "RemoveContainer" containerID="9c1a4f5d79ad64001bd9157a55394ac39bd51590bc1287dc6daa3348dab0dcd9" Oct 04 08:34:55 crc kubenswrapper[4969]: I1004 08:34:55.069110 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63362185-ca5e-442d-8684-ee9a819ddb9e" path="/var/lib/kubelet/pods/63362185-ca5e-442d-8684-ee9a819ddb9e/volumes" Oct 04 08:35:00 crc kubenswrapper[4969]: E1004 08:35:00.354669 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Oct 04 08:35:00 crc kubenswrapper[4969]: E1004 08:35:00.355327 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w6w9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(ff1b1453-89a7-4b5e-a55d-246e10fa18c2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 08:35:00 crc kubenswrapper[4969]: E1004 08:35:00.356881 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" Oct 04 08:35:00 crc kubenswrapper[4969]: I1004 08:35:00.544063 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="ceilometer-central-agent" containerID="cri-o://667efa089b3b8da7f77a98d0eec9a4f3d66b2d8f64cbb587b1e99170da6b37ab" gracePeriod=30 Oct 04 08:35:00 crc kubenswrapper[4969]: I1004 08:35:00.544125 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="ceilometer-notification-agent" containerID="cri-o://779b827573841530d0d9c1ad5778e8eb0addca0b6cf2069e39c769eb1953d8db" gracePeriod=30 Oct 04 08:35:00 crc kubenswrapper[4969]: I1004 08:35:00.544097 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="sg-core" containerID="cri-o://b080b21623eac49fc2e1208866892b6b46784834630fbd4864315ed3db91409a" gracePeriod=30 Oct 04 08:35:01 crc kubenswrapper[4969]: E1004 08:35:01.557513 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 04 08:35:01 crc kubenswrapper[4969]: E1004 08:35:01.557944 4969 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 04 08:35:01 crc kubenswrapper[4969]: E1004 08:35:01.558201 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.83:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-st8n9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-xjc7p_openstack(278ebf83-6c66-4d97-8b5c-a7a83ace00b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 08:35:01 crc kubenswrapper[4969]: E1004 08:35:01.562107 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-xjc7p" podUID="278ebf83-6c66-4d97-8b5c-a7a83ace00b7" Oct 04 08:35:01 crc kubenswrapper[4969]: I1004 08:35:01.568794 4969 scope.go:117] "RemoveContainer" containerID="3c69c1d4abd7a0fd6705c86e1214849291b2d5fe7e6e6a1d150d283f6dbfdc2a" Oct 04 08:35:01 crc kubenswrapper[4969]: I1004 08:35:01.577065 4969 generic.go:334] "Generic (PLEG): container finished" podID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerID="b080b21623eac49fc2e1208866892b6b46784834630fbd4864315ed3db91409a" exitCode=2 Oct 04 08:35:01 crc kubenswrapper[4969]: I1004 08:35:01.577099 4969 generic.go:334] "Generic (PLEG): container finished" podID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerID="667efa089b3b8da7f77a98d0eec9a4f3d66b2d8f64cbb587b1e99170da6b37ab" exitCode=0 Oct 04 08:35:01 crc kubenswrapper[4969]: I1004 08:35:01.577121 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff1b1453-89a7-4b5e-a55d-246e10fa18c2","Type":"ContainerDied","Data":"b080b21623eac49fc2e1208866892b6b46784834630fbd4864315ed3db91409a"} Oct 04 08:35:01 crc kubenswrapper[4969]: I1004 08:35:01.577147 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff1b1453-89a7-4b5e-a55d-246e10fa18c2","Type":"ContainerDied","Data":"667efa089b3b8da7f77a98d0eec9a4f3d66b2d8f64cbb587b1e99170da6b37ab"} Oct 04 08:35:01 crc kubenswrapper[4969]: I1004 08:35:01.670413 4969 scope.go:117] "RemoveContainer" containerID="b6d04d5b0f1a6714bbba1655346038197530bc75656bf81a4491925c3711b8ea" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.029332 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-569b746d9d-r7c45"] Oct 04 08:35:02 crc kubenswrapper[4969]: W1004 08:35:02.040821 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6425541_fb00_4545_a4df_64aba5d7e5ab.slice/crio-b85f538fd0f41045b5157e2a5e9d5b34cd015ca84bfcfcf0a0546a864d6769d0 WatchSource:0}: Error finding container b85f538fd0f41045b5157e2a5e9d5b34cd015ca84bfcfcf0a0546a864d6769d0: Status 404 returned error can't find the container with id b85f538fd0f41045b5157e2a5e9d5b34cd015ca84bfcfcf0a0546a864d6769d0 Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.117363 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.588757 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-569b746d9d-r7c45" event={"ID":"d6425541-fb00-4545-a4df-64aba5d7e5ab","Type":"ContainerStarted","Data":"a29f0f74d65caf8e02a4a41b87c4b9843e5e697a0763afb24cc2f5abdadb6147"} Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.589042 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-569b746d9d-r7c45" event={"ID":"d6425541-fb00-4545-a4df-64aba5d7e5ab","Type":"ContainerStarted","Data":"f4825e5ddd5a8342e76e76c7b020ab6b076e65383cd61308318337e93fe6e2e0"} Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.589057 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-569b746d9d-r7c45" event={"ID":"d6425541-fb00-4545-a4df-64aba5d7e5ab","Type":"ContainerStarted","Data":"b85f538fd0f41045b5157e2a5e9d5b34cd015ca84bfcfcf0a0546a864d6769d0"} Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.590571 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.590624 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.605674 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ktnkt" event={"ID":"b1bc4bef-f750-4b65-96c5-af742379cf87","Type":"ContainerStarted","Data":"4125e33fb697f71057aa46cf79725094aaf92fada934f2163ea811c16b4176ea"} Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.609502 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerStarted","Data":"a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6"} Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.622407 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-569b746d9d-r7c45" podStartSLOduration=12.622385473 podStartE2EDuration="12.622385473s" podCreationTimestamp="2025-10-04 08:34:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:02.610094643 +0000 UTC m=+1130.364363457" watchObservedRunningTime="2025-10-04 08:35:02.622385473 +0000 UTC m=+1130.376654287" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.622970 4969 generic.go:334] "Generic (PLEG): container finished" podID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerID="779b827573841530d0d9c1ad5778e8eb0addca0b6cf2069e39c769eb1953d8db" exitCode=0 Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.623046 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff1b1453-89a7-4b5e-a55d-246e10fa18c2","Type":"ContainerDied","Data":"779b827573841530d0d9c1ad5778e8eb0addca0b6cf2069e39c769eb1953d8db"} Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.624945 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"b36dc326-a420-42e1-a274-9e82641ee148","Type":"ContainerStarted","Data":"9797dce8b1b373c587c9a6daf0d8d5ac3d1600c03698b249717c6d9937300cff"} Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.624971 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"b36dc326-a420-42e1-a274-9e82641ee148","Type":"ContainerStarted","Data":"9222af113dd9551fbfa3d59704b5a3ec1b99d9370e8fa72f47cf5dbcba16376b"} Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.624982 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"b36dc326-a420-42e1-a274-9e82641ee148","Type":"ContainerStarted","Data":"78b22738ffaba69b04ce23545b90f71ec2d023fa8082df6d61a76215b6e1a8bc"} Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.625229 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 04 08:35:02 crc kubenswrapper[4969]: E1004 08:35:02.626768 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.83:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-xjc7p" podUID="278ebf83-6c66-4d97-8b5c-a7a83ace00b7" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.627916 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="b36dc326-a420-42e1-a274-9e82641ee148" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.169:9322/\": dial tcp 10.217.0.169:9322: connect: connection refused" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.662577 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ktnkt" podStartSLOduration=4.941383791 podStartE2EDuration="31.662561784s" podCreationTimestamp="2025-10-04 08:34:31 +0000 UTC" firstStartedPulling="2025-10-04 08:34:34.823354457 +0000 UTC m=+1102.577623271" lastFinishedPulling="2025-10-04 08:35:01.54453244 +0000 UTC m=+1129.298801264" observedRunningTime="2025-10-04 08:35:02.654294252 +0000 UTC m=+1130.408563076" watchObservedRunningTime="2025-10-04 08:35:02.662561784 +0000 UTC m=+1130.416830598" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.801495 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.825782 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=13.825763458 podStartE2EDuration="13.825763458s" podCreationTimestamp="2025-10-04 08:34:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:02.688905807 +0000 UTC m=+1130.443174631" watchObservedRunningTime="2025-10-04 08:35:02.825763458 +0000 UTC m=+1130.580032272" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.890463 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-combined-ca-bundle\") pod \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.890599 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-log-httpd\") pod \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.890662 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-scripts\") pod \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.890720 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-config-data\") pod \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.890746 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-sg-core-conf-yaml\") pod \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.890807 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w6w9\" (UniqueName: \"kubernetes.io/projected/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-kube-api-access-8w6w9\") pod \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.890922 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-run-httpd\") pod \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\" (UID: \"ff1b1453-89a7-4b5e-a55d-246e10fa18c2\") " Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.891547 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ff1b1453-89a7-4b5e-a55d-246e10fa18c2" (UID: "ff1b1453-89a7-4b5e-a55d-246e10fa18c2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.891765 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ff1b1453-89a7-4b5e-a55d-246e10fa18c2" (UID: "ff1b1453-89a7-4b5e-a55d-246e10fa18c2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.897465 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-kube-api-access-8w6w9" (OuterVolumeSpecName: "kube-api-access-8w6w9") pod "ff1b1453-89a7-4b5e-a55d-246e10fa18c2" (UID: "ff1b1453-89a7-4b5e-a55d-246e10fa18c2"). InnerVolumeSpecName "kube-api-access-8w6w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.904432 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-scripts" (OuterVolumeSpecName: "scripts") pod "ff1b1453-89a7-4b5e-a55d-246e10fa18c2" (UID: "ff1b1453-89a7-4b5e-a55d-246e10fa18c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.964004 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff1b1453-89a7-4b5e-a55d-246e10fa18c2" (UID: "ff1b1453-89a7-4b5e-a55d-246e10fa18c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.971671 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ff1b1453-89a7-4b5e-a55d-246e10fa18c2" (UID: "ff1b1453-89a7-4b5e-a55d-246e10fa18c2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.990695 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-config-data" (OuterVolumeSpecName: "config-data") pod "ff1b1453-89a7-4b5e-a55d-246e10fa18c2" (UID: "ff1b1453-89a7-4b5e-a55d-246e10fa18c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.997394 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.998579 4969 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.998811 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.998916 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.999007 4969 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.999100 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w6w9\" (UniqueName: \"kubernetes.io/projected/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-kube-api-access-8w6w9\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:02 crc kubenswrapper[4969]: I1004 08:35:02.999191 4969 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff1b1453-89a7-4b5e-a55d-246e10fa18c2-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.661182 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.662161 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff1b1453-89a7-4b5e-a55d-246e10fa18c2","Type":"ContainerDied","Data":"8eb03613ffbe4b3fc758b52cd1d456d72cfc2842c4467052b751aebbfa1e0318"} Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.662208 4969 scope.go:117] "RemoveContainer" containerID="b080b21623eac49fc2e1208866892b6b46784834630fbd4864315ed3db91409a" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.715413 4969 scope.go:117] "RemoveContainer" containerID="779b827573841530d0d9c1ad5778e8eb0addca0b6cf2069e39c769eb1953d8db" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.778852 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.782221 4969 scope.go:117] "RemoveContainer" containerID="667efa089b3b8da7f77a98d0eec9a4f3d66b2d8f64cbb587b1e99170da6b37ab" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.799614 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.810778 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:03 crc kubenswrapper[4969]: E1004 08:35:03.811265 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="ceilometer-notification-agent" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.811289 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="ceilometer-notification-agent" Oct 04 08:35:03 crc kubenswrapper[4969]: E1004 08:35:03.811300 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="ceilometer-central-agent" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.811309 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="ceilometer-central-agent" Oct 04 08:35:03 crc kubenswrapper[4969]: E1004 08:35:03.811334 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="sg-core" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.811342 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="sg-core" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.811665 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="sg-core" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.811695 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="ceilometer-notification-agent" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.811705 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" containerName="ceilometer-central-agent" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.813838 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.817193 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.817407 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.821756 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.914730 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-scripts\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.914777 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.914814 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.914850 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gl4p\" (UniqueName: \"kubernetes.io/projected/56d85928-e2de-4594-8578-b504979d67fc-kube-api-access-6gl4p\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.914890 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-config-data\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.914916 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-run-httpd\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:03 crc kubenswrapper[4969]: I1004 08:35:03.914933 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-log-httpd\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.016865 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gl4p\" (UniqueName: \"kubernetes.io/projected/56d85928-e2de-4594-8578-b504979d67fc-kube-api-access-6gl4p\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.017283 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-config-data\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.018695 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-run-httpd\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.019084 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-log-httpd\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.019444 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-run-httpd\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.019595 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-log-httpd\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.019775 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-scripts\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.019885 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.019977 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.027049 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-config-data\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.027714 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.034141 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.035568 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-scripts\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.040150 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gl4p\" (UniqueName: \"kubernetes.io/projected/56d85928-e2de-4594-8578-b504979d67fc-kube-api-access-6gl4p\") pod \"ceilometer-0\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.140727 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.426908 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-586c975586-mxhpd" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.622712 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:35:04 crc kubenswrapper[4969]: E1004 08:35:04.623483 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6 is running failed: container process not found" containerID="a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Oct 04 08:35:04 crc kubenswrapper[4969]: E1004 08:35:04.625058 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6 is running failed: container process not found" containerID="a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Oct 04 08:35:04 crc kubenswrapper[4969]: E1004 08:35:04.625804 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6 is running failed: container process not found" containerID="a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Oct 04 08:35:04 crc kubenswrapper[4969]: E1004 08:35:04.625891 4969 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6 is running failed: container process not found" probeType="Startup" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.682110 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.687608 4969 generic.go:334] "Generic (PLEG): container finished" podID="7144eb34-59d2-4583-8c46-3e56258966da" containerID="a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6" exitCode=1 Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.687696 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerDied","Data":"a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6"} Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.687914 4969 scope.go:117] "RemoveContainer" containerID="9c1a4f5d79ad64001bd9157a55394ac39bd51590bc1287dc6daa3348dab0dcd9" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.689046 4969 scope.go:117] "RemoveContainer" containerID="a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6" Oct 04 08:35:04 crc kubenswrapper[4969]: E1004 08:35:04.690145 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:35:04 crc kubenswrapper[4969]: I1004 08:35:04.866464 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 04 08:35:05 crc kubenswrapper[4969]: I1004 08:35:05.069367 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff1b1453-89a7-4b5e-a55d-246e10fa18c2" path="/var/lib/kubelet/pods/ff1b1453-89a7-4b5e-a55d-246e10fa18c2/volumes" Oct 04 08:35:05 crc kubenswrapper[4969]: I1004 08:35:05.728532 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerStarted","Data":"6cc5b7b41370dedd9ec0b0a195bc591aa46868845bb4ecdda84c51c080a6e195"} Oct 04 08:35:05 crc kubenswrapper[4969]: I1004 08:35:05.728945 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerStarted","Data":"b768e5ff863031298333e33ac095d23331b27d4256926c232249e752b9760ca0"} Oct 04 08:35:05 crc kubenswrapper[4969]: I1004 08:35:05.728964 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerStarted","Data":"e8b745d7981bf7bcf73e64eada07769197c859be75053d9131ee19c91c66db53"} Oct 04 08:35:05 crc kubenswrapper[4969]: I1004 08:35:05.731741 4969 scope.go:117] "RemoveContainer" containerID="a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6" Oct 04 08:35:05 crc kubenswrapper[4969]: E1004 08:35:05.732102 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:35:05 crc kubenswrapper[4969]: I1004 08:35:05.940067 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 04 08:35:06 crc kubenswrapper[4969]: I1004 08:35:06.746069 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerStarted","Data":"4974cbb3139927f9e1a46aea06afe4cb091e883ecef931f80ed9502a336d98f0"} Oct 04 08:35:06 crc kubenswrapper[4969]: I1004 08:35:06.748404 4969 generic.go:334] "Generic (PLEG): container finished" podID="b1bc4bef-f750-4b65-96c5-af742379cf87" containerID="4125e33fb697f71057aa46cf79725094aaf92fada934f2163ea811c16b4176ea" exitCode=0 Oct 04 08:35:06 crc kubenswrapper[4969]: I1004 08:35:06.748452 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ktnkt" event={"ID":"b1bc4bef-f750-4b65-96c5-af742379cf87","Type":"ContainerDied","Data":"4125e33fb697f71057aa46cf79725094aaf92fada934f2163ea811c16b4176ea"} Oct 04 08:35:07 crc kubenswrapper[4969]: I1004 08:35:07.764753 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerStarted","Data":"25989098a2aa6cbda783ccefe89d7b2eb1ca44bb80e94c75353ebb8630868e1d"} Oct 04 08:35:07 crc kubenswrapper[4969]: I1004 08:35:07.765133 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 08:35:07 crc kubenswrapper[4969]: I1004 08:35:07.797664 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.466222529 podStartE2EDuration="4.797641635s" podCreationTimestamp="2025-10-04 08:35:03 +0000 UTC" firstStartedPulling="2025-10-04 08:35:04.725145797 +0000 UTC m=+1132.479414621" lastFinishedPulling="2025-10-04 08:35:07.056564873 +0000 UTC m=+1134.810833727" observedRunningTime="2025-10-04 08:35:07.796800365 +0000 UTC m=+1135.551069219" watchObservedRunningTime="2025-10-04 08:35:07.797641635 +0000 UTC m=+1135.551910459" Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.128847 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.216990 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcdtk\" (UniqueName: \"kubernetes.io/projected/b1bc4bef-f750-4b65-96c5-af742379cf87-kube-api-access-wcdtk\") pod \"b1bc4bef-f750-4b65-96c5-af742379cf87\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.217104 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-combined-ca-bundle\") pod \"b1bc4bef-f750-4b65-96c5-af742379cf87\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.217249 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-db-sync-config-data\") pod \"b1bc4bef-f750-4b65-96c5-af742379cf87\" (UID: \"b1bc4bef-f750-4b65-96c5-af742379cf87\") " Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.226555 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1bc4bef-f750-4b65-96c5-af742379cf87-kube-api-access-wcdtk" (OuterVolumeSpecName: "kube-api-access-wcdtk") pod "b1bc4bef-f750-4b65-96c5-af742379cf87" (UID: "b1bc4bef-f750-4b65-96c5-af742379cf87"). InnerVolumeSpecName "kube-api-access-wcdtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.235510 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b1bc4bef-f750-4b65-96c5-af742379cf87" (UID: "b1bc4bef-f750-4b65-96c5-af742379cf87"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.257124 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1bc4bef-f750-4b65-96c5-af742379cf87" (UID: "b1bc4bef-f750-4b65-96c5-af742379cf87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.319684 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.319714 4969 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1bc4bef-f750-4b65-96c5-af742379cf87-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.319725 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcdtk\" (UniqueName: \"kubernetes.io/projected/b1bc4bef-f750-4b65-96c5-af742379cf87-kube-api-access-wcdtk\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.776128 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ktnkt" event={"ID":"b1bc4bef-f750-4b65-96c5-af742379cf87","Type":"ContainerDied","Data":"098dc0c3b31cc252000abccd87dff62b330a236a61ef2f28d1c5463c342022b4"} Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.776206 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="098dc0c3b31cc252000abccd87dff62b330a236a61ef2f28d1c5463c342022b4" Oct 04 08:35:08 crc kubenswrapper[4969]: I1004 08:35:08.776166 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ktnkt" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.099589 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-9fb8f975c-4l4lz"] Oct 04 08:35:09 crc kubenswrapper[4969]: E1004 08:35:09.100286 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1bc4bef-f750-4b65-96c5-af742379cf87" containerName="barbican-db-sync" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.100327 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1bc4bef-f750-4b65-96c5-af742379cf87" containerName="barbican-db-sync" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.100776 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1bc4bef-f750-4b65-96c5-af742379cf87" containerName="barbican-db-sync" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.102933 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.109060 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.109117 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8dgl8" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.109537 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.137482 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9fb8f975c-4l4lz"] Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.143387 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-logs\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.143479 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-combined-ca-bundle\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.143618 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-config-data-custom\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.143676 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkkvt\" (UniqueName: \"kubernetes.io/projected/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-kube-api-access-wkkvt\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.143706 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-config-data\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.156500 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c"] Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.158357 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.163889 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.181777 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c"] Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.237333 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d5ff7c687-t2ln2"] Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.238825 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245217 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm556\" (UniqueName: \"kubernetes.io/projected/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-kube-api-access-zm556\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245277 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-logs\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245373 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-logs\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245434 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-combined-ca-bundle\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245489 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-combined-ca-bundle\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245584 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-config-data\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245621 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-config-data-custom\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245661 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-config-data-custom\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245706 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkkvt\" (UniqueName: \"kubernetes.io/projected/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-kube-api-access-wkkvt\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245749 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-config-data\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.245818 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-logs\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.259064 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-combined-ca-bundle\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.259994 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-config-data-custom\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.261769 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-config-data\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.277950 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkkvt\" (UniqueName: \"kubernetes.io/projected/3f74b083-ba83-4f80-a8a6-e5e10e6dde28-kube-api-access-wkkvt\") pod \"barbican-worker-9fb8f975c-4l4lz\" (UID: \"3f74b083-ba83-4f80-a8a6-e5e10e6dde28\") " pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.333568 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d5ff7c687-t2ln2"] Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.347681 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-svc\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.347729 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-config-data-custom\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.347805 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm556\" (UniqueName: \"kubernetes.io/projected/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-kube-api-access-zm556\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.347823 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-logs\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.347868 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-nb\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.347902 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-swift-storage-0\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.347937 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-combined-ca-bundle\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.347960 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-config\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.347999 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-sb\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.348031 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vk5w\" (UniqueName: \"kubernetes.io/projected/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-kube-api-access-5vk5w\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.348064 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-config-data\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.355143 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-logs\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.376235 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-combined-ca-bundle\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.381063 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-config-data\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.381265 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-config-data-custom\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.392611 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm556\" (UniqueName: \"kubernetes.io/projected/f2e7b743-71f9-46c8-aeb7-450c63bf22ca-kube-api-access-zm556\") pod \"barbican-keystone-listener-7d5b5f7c58-kgs6c\" (UID: \"f2e7b743-71f9-46c8-aeb7-450c63bf22ca\") " pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.396572 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-798dcfcb48-rr6vl"] Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.399062 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.413846 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.423609 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9fb8f975c-4l4lz" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.426611 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-798dcfcb48-rr6vl"] Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.435855 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-699ff78d6b-dv46q" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467110 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-swift-storage-0\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467219 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-config\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467256 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d580e7-abf5-476a-8aaa-09a8415193e3-logs\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467293 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-sb\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467337 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-combined-ca-bundle\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467356 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vk5w\" (UniqueName: \"kubernetes.io/projected/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-kube-api-access-5vk5w\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467384 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467456 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data-custom\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467487 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-svc\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467588 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5hzl\" (UniqueName: \"kubernetes.io/projected/67d580e7-abf5-476a-8aaa-09a8415193e3-kube-api-access-b5hzl\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.467704 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-nb\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.468527 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-nb\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.469043 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-swift-storage-0\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.469571 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-config\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.470057 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-sb\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.470811 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-svc\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.490200 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.490393 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vk5w\" (UniqueName: \"kubernetes.io/projected/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-kube-api-access-5vk5w\") pod \"dnsmasq-dns-d5ff7c687-t2ln2\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.569826 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-combined-ca-bundle\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.570124 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.570156 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data-custom\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.570233 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5hzl\" (UniqueName: \"kubernetes.io/projected/67d580e7-abf5-476a-8aaa-09a8415193e3-kube-api-access-b5hzl\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.570388 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d580e7-abf5-476a-8aaa-09a8415193e3-logs\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.570818 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d580e7-abf5-476a-8aaa-09a8415193e3-logs\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.577536 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-combined-ca-bundle\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.586227 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data-custom\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.587061 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.593946 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5hzl\" (UniqueName: \"kubernetes.io/projected/67d580e7-abf5-476a-8aaa-09a8415193e3-kube-api-access-b5hzl\") pod \"barbican-api-798dcfcb48-rr6vl\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.650000 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.805076 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.866625 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.884184 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 04 08:35:09 crc kubenswrapper[4969]: I1004 08:35:09.939268 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9fb8f975c-4l4lz"] Oct 04 08:35:10 crc kubenswrapper[4969]: I1004 08:35:10.090485 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c"] Oct 04 08:35:10 crc kubenswrapper[4969]: I1004 08:35:10.229176 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d5ff7c687-t2ln2"] Oct 04 08:35:10 crc kubenswrapper[4969]: W1004 08:35:10.332125 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67d580e7_abf5_476a_8aaa_09a8415193e3.slice/crio-9d9da6d45f95742828c1c159963971908f4992faa1215e58c25729f7679aa132 WatchSource:0}: Error finding container 9d9da6d45f95742828c1c159963971908f4992faa1215e58c25729f7679aa132: Status 404 returned error can't find the container with id 9d9da6d45f95742828c1c159963971908f4992faa1215e58c25729f7679aa132 Oct 04 08:35:10 crc kubenswrapper[4969]: I1004 08:35:10.334793 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-798dcfcb48-rr6vl"] Oct 04 08:35:10 crc kubenswrapper[4969]: I1004 08:35:10.806596 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" event={"ID":"aaf3be84-c9a9-481c-8e10-a5dc928f9be9","Type":"ContainerStarted","Data":"680b10407f930ea70abcfd5bdcd4452f5afd5516364e2b64037c50e6c4407176"} Oct 04 08:35:10 crc kubenswrapper[4969]: I1004 08:35:10.807712 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" event={"ID":"f2e7b743-71f9-46c8-aeb7-450c63bf22ca","Type":"ContainerStarted","Data":"18c4b3c808a86536085a4b41be0635db90ae5b246f49303484ad47596635c354"} Oct 04 08:35:10 crc kubenswrapper[4969]: I1004 08:35:10.809126 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798dcfcb48-rr6vl" event={"ID":"67d580e7-abf5-476a-8aaa-09a8415193e3","Type":"ContainerStarted","Data":"9d9da6d45f95742828c1c159963971908f4992faa1215e58c25729f7679aa132"} Oct 04 08:35:10 crc kubenswrapper[4969]: I1004 08:35:10.810169 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9fb8f975c-4l4lz" event={"ID":"3f74b083-ba83-4f80-a8a6-e5e10e6dde28","Type":"ContainerStarted","Data":"06c70b14b20eb6a121c24d2f4a5a10f3f1c176e47fdb1d9704ef71d99f5910a4"} Oct 04 08:35:10 crc kubenswrapper[4969]: I1004 08:35:10.819746 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.463653 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.466662 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.470946 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.471146 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-8nqvf" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.471261 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.499068 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.533554 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5532f51-e305-4741-8661-25029faf98f2-openstack-config-secret\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.533745 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5532f51-e305-4741-8661-25029faf98f2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.533779 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5532f51-e305-4741-8661-25029faf98f2-openstack-config\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.533866 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzccb\" (UniqueName: \"kubernetes.io/projected/f5532f51-e305-4741-8661-25029faf98f2-kube-api-access-dzccb\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.639661 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzccb\" (UniqueName: \"kubernetes.io/projected/f5532f51-e305-4741-8661-25029faf98f2-kube-api-access-dzccb\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.639848 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5532f51-e305-4741-8661-25029faf98f2-openstack-config-secret\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.639934 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5532f51-e305-4741-8661-25029faf98f2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.639963 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5532f51-e305-4741-8661-25029faf98f2-openstack-config\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.655944 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5532f51-e305-4741-8661-25029faf98f2-openstack-config\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.668442 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5532f51-e305-4741-8661-25029faf98f2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.675569 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5532f51-e305-4741-8661-25029faf98f2-openstack-config-secret\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.679054 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzccb\" (UniqueName: \"kubernetes.io/projected/f5532f51-e305-4741-8661-25029faf98f2-kube-api-access-dzccb\") pod \"openstackclient\" (UID: \"f5532f51-e305-4741-8661-25029faf98f2\") " pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.834455 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798dcfcb48-rr6vl" event={"ID":"67d580e7-abf5-476a-8aaa-09a8415193e3","Type":"ContainerStarted","Data":"569dd599a2462b73e18ddf2712ae36e6399c2b7b176b25bbd0a2393a5d7966cb"} Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.834513 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798dcfcb48-rr6vl" event={"ID":"67d580e7-abf5-476a-8aaa-09a8415193e3","Type":"ContainerStarted","Data":"7e4ef3e6610aa4c53eca0a0bf956e1000c461ac0ca56693894ae8703622d3b98"} Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.834534 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.834548 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.839372 4969 generic.go:334] "Generic (PLEG): container finished" podID="aaf3be84-c9a9-481c-8e10-a5dc928f9be9" containerID="989d9e3c86ede525cc5ffde902df95e5ef1fff7290f1cf9aac0416ee0bede741" exitCode=0 Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.839456 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" event={"ID":"aaf3be84-c9a9-481c-8e10-a5dc928f9be9","Type":"ContainerDied","Data":"989d9e3c86ede525cc5ffde902df95e5ef1fff7290f1cf9aac0416ee0bede741"} Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.845482 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 08:35:11 crc kubenswrapper[4969]: I1004 08:35:11.866538 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-798dcfcb48-rr6vl" podStartSLOduration=2.866518187 podStartE2EDuration="2.866518187s" podCreationTimestamp="2025-10-04 08:35:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:11.854622627 +0000 UTC m=+1139.608891461" watchObservedRunningTime="2025-10-04 08:35:11.866518187 +0000 UTC m=+1139.620787001" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.206045 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6dbff44956-qlvk2"] Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.216496 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.224538 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.234056 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.239057 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dbff44956-qlvk2"] Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.255435 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-public-tls-certs\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.255486 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-combined-ca-bundle\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.255533 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d46c925d-1b2f-4696-b099-f3a549dda18b-logs\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.255556 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-internal-tls-certs\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.255578 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-config-data-custom\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.255612 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78g29\" (UniqueName: \"kubernetes.io/projected/d46c925d-1b2f-4696-b099-f3a549dda18b-kube-api-access-78g29\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.255655 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-config-data\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.357145 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-public-tls-certs\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.357197 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-combined-ca-bundle\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.357248 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d46c925d-1b2f-4696-b099-f3a549dda18b-logs\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.357275 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-internal-tls-certs\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.357298 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-config-data-custom\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.357339 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78g29\" (UniqueName: \"kubernetes.io/projected/d46c925d-1b2f-4696-b099-f3a549dda18b-kube-api-access-78g29\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.357382 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-config-data\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.358321 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d46c925d-1b2f-4696-b099-f3a549dda18b-logs\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.362100 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-internal-tls-certs\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.362231 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-config-data\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.384550 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-public-tls-certs\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.385474 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-config-data-custom\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.386665 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46c925d-1b2f-4696-b099-f3a549dda18b-combined-ca-bundle\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.391251 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78g29\" (UniqueName: \"kubernetes.io/projected/d46c925d-1b2f-4696-b099-f3a549dda18b-kube-api-access-78g29\") pod \"barbican-api-6dbff44956-qlvk2\" (UID: \"d46c925d-1b2f-4696-b099-f3a549dda18b\") " pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:12 crc kubenswrapper[4969]: I1004 08:35:12.637211 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.376127 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 08:35:13 crc kubenswrapper[4969]: W1004 08:35:13.381464 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5532f51_e305_4741_8661_25029faf98f2.slice/crio-b562208d6967b175416e3a9559e69518626ffcc46901a72ce4cd46492bd8f303 WatchSource:0}: Error finding container b562208d6967b175416e3a9559e69518626ffcc46901a72ce4cd46492bd8f303: Status 404 returned error can't find the container with id b562208d6967b175416e3a9559e69518626ffcc46901a72ce4cd46492bd8f303 Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.398920 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dbff44956-qlvk2"] Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.880406 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dbff44956-qlvk2" event={"ID":"d46c925d-1b2f-4696-b099-f3a549dda18b","Type":"ContainerStarted","Data":"57d7ce2f614f42e56ccb6ebc7322e9fe086e44d69e51655f3d87395bc4de80c7"} Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.880565 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dbff44956-qlvk2" event={"ID":"d46c925d-1b2f-4696-b099-f3a549dda18b","Type":"ContainerStarted","Data":"d654369e3916fa04efee3500804e9b7075422d28b937b0529c586a515b6c280e"} Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.883261 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9fb8f975c-4l4lz" event={"ID":"3f74b083-ba83-4f80-a8a6-e5e10e6dde28","Type":"ContainerStarted","Data":"33424c32b52fcdc6bf6f7a241e414c8446ac9d96fc9eaa4b0a7f71fb5035cfe3"} Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.883306 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9fb8f975c-4l4lz" event={"ID":"3f74b083-ba83-4f80-a8a6-e5e10e6dde28","Type":"ContainerStarted","Data":"cc497836ef7275744e7f070af698c23274627e74959be99d9338bb59e688843f"} Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.884806 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f5532f51-e305-4741-8661-25029faf98f2","Type":"ContainerStarted","Data":"b562208d6967b175416e3a9559e69518626ffcc46901a72ce4cd46492bd8f303"} Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.890016 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" event={"ID":"aaf3be84-c9a9-481c-8e10-a5dc928f9be9","Type":"ContainerStarted","Data":"8163280de698d77ca2a24a53653e964d04932b8276ae73466ee2a11a1103a85c"} Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.890170 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.892148 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" event={"ID":"f2e7b743-71f9-46c8-aeb7-450c63bf22ca","Type":"ContainerStarted","Data":"35e4ea8de8eb2355a767312c6283808ac14220c42fa689225189186a592f5c84"} Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.892479 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" event={"ID":"f2e7b743-71f9-46c8-aeb7-450c63bf22ca","Type":"ContainerStarted","Data":"b920e7864744a35b1e700e0f64470066170332ba3fcadbbc845080b60ee8997f"} Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.904040 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-9fb8f975c-4l4lz" podStartSLOduration=2.022062671 podStartE2EDuration="4.904019967s" podCreationTimestamp="2025-10-04 08:35:09 +0000 UTC" firstStartedPulling="2025-10-04 08:35:09.975780759 +0000 UTC m=+1137.730049563" lastFinishedPulling="2025-10-04 08:35:12.857738045 +0000 UTC m=+1140.612006859" observedRunningTime="2025-10-04 08:35:13.902340606 +0000 UTC m=+1141.656609420" watchObservedRunningTime="2025-10-04 08:35:13.904019967 +0000 UTC m=+1141.658288781" Oct 04 08:35:13 crc kubenswrapper[4969]: I1004 08:35:13.926749 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7d5b5f7c58-kgs6c" podStartSLOduration=2.139531139 podStartE2EDuration="4.926730712s" podCreationTimestamp="2025-10-04 08:35:09 +0000 UTC" firstStartedPulling="2025-10-04 08:35:10.083438997 +0000 UTC m=+1137.837707811" lastFinishedPulling="2025-10-04 08:35:12.87063857 +0000 UTC m=+1140.624907384" observedRunningTime="2025-10-04 08:35:13.922633232 +0000 UTC m=+1141.676902046" watchObservedRunningTime="2025-10-04 08:35:13.926730712 +0000 UTC m=+1141.680999526" Oct 04 08:35:14 crc kubenswrapper[4969]: I1004 08:35:14.428398 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-586c975586-mxhpd" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Oct 04 08:35:14 crc kubenswrapper[4969]: I1004 08:35:14.428578 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:35:14 crc kubenswrapper[4969]: I1004 08:35:14.452521 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" podStartSLOduration=5.452496827 podStartE2EDuration="5.452496827s" podCreationTimestamp="2025-10-04 08:35:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:13.949115709 +0000 UTC m=+1141.703384543" watchObservedRunningTime="2025-10-04 08:35:14.452496827 +0000 UTC m=+1142.206765641" Oct 04 08:35:14 crc kubenswrapper[4969]: I1004 08:35:14.622913 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:35:14 crc kubenswrapper[4969]: I1004 08:35:14.623694 4969 scope.go:117] "RemoveContainer" containerID="a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6" Oct 04 08:35:14 crc kubenswrapper[4969]: E1004 08:35:14.624092 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:35:14 crc kubenswrapper[4969]: I1004 08:35:14.904983 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dbff44956-qlvk2" event={"ID":"d46c925d-1b2f-4696-b099-f3a549dda18b","Type":"ContainerStarted","Data":"264fa3e4c046bace3bd03d1ef0e7cc526ea299b5efa82f403a59d1fb9db96f7b"} Oct 04 08:35:14 crc kubenswrapper[4969]: I1004 08:35:14.905896 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:14 crc kubenswrapper[4969]: I1004 08:35:14.930484 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6dbff44956-qlvk2" podStartSLOduration=2.930412745 podStartE2EDuration="2.930412745s" podCreationTimestamp="2025-10-04 08:35:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:14.922285296 +0000 UTC m=+1142.676554170" watchObservedRunningTime="2025-10-04 08:35:14.930412745 +0000 UTC m=+1142.684681599" Oct 04 08:35:15 crc kubenswrapper[4969]: I1004 08:35:15.922750 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.075441 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-75f87ddd45-2md8q"] Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.077038 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.079379 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.079933 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.080006 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.083403 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-75f87ddd45-2md8q"] Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.140891 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2887852-0cd7-476e-8c4e-1ed98c66ede4-log-httpd\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.140937 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5njpr\" (UniqueName: \"kubernetes.io/projected/f2887852-0cd7-476e-8c4e-1ed98c66ede4-kube-api-access-5njpr\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.140970 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2887852-0cd7-476e-8c4e-1ed98c66ede4-run-httpd\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.140993 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-combined-ca-bundle\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.141030 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-public-tls-certs\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.141081 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f2887852-0cd7-476e-8c4e-1ed98c66ede4-etc-swift\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.141118 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-config-data\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.141252 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-internal-tls-certs\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.243439 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2887852-0cd7-476e-8c4e-1ed98c66ede4-log-httpd\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.243515 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5njpr\" (UniqueName: \"kubernetes.io/projected/f2887852-0cd7-476e-8c4e-1ed98c66ede4-kube-api-access-5njpr\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.243539 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2887852-0cd7-476e-8c4e-1ed98c66ede4-run-httpd\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.243557 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-combined-ca-bundle\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.243599 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-public-tls-certs\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.243631 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f2887852-0cd7-476e-8c4e-1ed98c66ede4-etc-swift\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.243670 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-config-data\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.243720 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-internal-tls-certs\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.243852 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2887852-0cd7-476e-8c4e-1ed98c66ede4-log-httpd\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.245690 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2887852-0cd7-476e-8c4e-1ed98c66ede4-run-httpd\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.250038 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-public-tls-certs\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.255184 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-internal-tls-certs\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.255513 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f2887852-0cd7-476e-8c4e-1ed98c66ede4-etc-swift\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.256148 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-config-data\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.257937 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2887852-0cd7-476e-8c4e-1ed98c66ede4-combined-ca-bundle\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.268268 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5njpr\" (UniqueName: \"kubernetes.io/projected/f2887852-0cd7-476e-8c4e-1ed98c66ede4-kube-api-access-5njpr\") pod \"swift-proxy-75f87ddd45-2md8q\" (UID: \"f2887852-0cd7-476e-8c4e-1ed98c66ede4\") " pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.431039 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.934921 4969 generic.go:334] "Generic (PLEG): container finished" podID="2861dd9f-5291-4f2a-bb30-0b9c54baa5cf" containerID="fa7d3b44e484120a5a3ea50089d66bad7288ad199d49aca435555dc55b054f41" exitCode=0 Oct 04 08:35:16 crc kubenswrapper[4969]: I1004 08:35:16.935509 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cf9ls" event={"ID":"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf","Type":"ContainerDied","Data":"fa7d3b44e484120a5a3ea50089d66bad7288ad199d49aca435555dc55b054f41"} Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.039512 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-75f87ddd45-2md8q"] Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.164268 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-bx4tb"] Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.165473 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bx4tb" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.177817 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bx4tb"] Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.255991 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-7fphd"] Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.257245 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7fphd" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.265501 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s54n\" (UniqueName: \"kubernetes.io/projected/f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f-kube-api-access-4s54n\") pod \"nova-api-db-create-bx4tb\" (UID: \"f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f\") " pod="openstack/nova-api-db-create-bx4tb" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.273118 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-7fphd"] Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.366803 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s54n\" (UniqueName: \"kubernetes.io/projected/f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f-kube-api-access-4s54n\") pod \"nova-api-db-create-bx4tb\" (UID: \"f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f\") " pod="openstack/nova-api-db-create-bx4tb" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.366931 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfzlj\" (UniqueName: \"kubernetes.io/projected/1ff3f16f-8103-49a7-8a5c-07b6b7f399e7-kube-api-access-pfzlj\") pod \"nova-cell0-db-create-7fphd\" (UID: \"1ff3f16f-8103-49a7-8a5c-07b6b7f399e7\") " pod="openstack/nova-cell0-db-create-7fphd" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.370196 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-wjg5p"] Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.371635 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wjg5p" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.378308 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wjg5p"] Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.407449 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s54n\" (UniqueName: \"kubernetes.io/projected/f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f-kube-api-access-4s54n\") pod \"nova-api-db-create-bx4tb\" (UID: \"f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f\") " pod="openstack/nova-api-db-create-bx4tb" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.468082 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfzlj\" (UniqueName: \"kubernetes.io/projected/1ff3f16f-8103-49a7-8a5c-07b6b7f399e7-kube-api-access-pfzlj\") pod \"nova-cell0-db-create-7fphd\" (UID: \"1ff3f16f-8103-49a7-8a5c-07b6b7f399e7\") " pod="openstack/nova-cell0-db-create-7fphd" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.468197 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnp9p\" (UniqueName: \"kubernetes.io/projected/46a217c4-ea44-4608-9300-65410bec2b88-kube-api-access-dnp9p\") pod \"nova-cell1-db-create-wjg5p\" (UID: \"46a217c4-ea44-4608-9300-65410bec2b88\") " pod="openstack/nova-cell1-db-create-wjg5p" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.480890 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bx4tb" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.490039 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfzlj\" (UniqueName: \"kubernetes.io/projected/1ff3f16f-8103-49a7-8a5c-07b6b7f399e7-kube-api-access-pfzlj\") pod \"nova-cell0-db-create-7fphd\" (UID: \"1ff3f16f-8103-49a7-8a5c-07b6b7f399e7\") " pod="openstack/nova-cell0-db-create-7fphd" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.570209 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnp9p\" (UniqueName: \"kubernetes.io/projected/46a217c4-ea44-4608-9300-65410bec2b88-kube-api-access-dnp9p\") pod \"nova-cell1-db-create-wjg5p\" (UID: \"46a217c4-ea44-4608-9300-65410bec2b88\") " pod="openstack/nova-cell1-db-create-wjg5p" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.582625 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7fphd" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.586876 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnp9p\" (UniqueName: \"kubernetes.io/projected/46a217c4-ea44-4608-9300-65410bec2b88-kube-api-access-dnp9p\") pod \"nova-cell1-db-create-wjg5p\" (UID: \"46a217c4-ea44-4608-9300-65410bec2b88\") " pod="openstack/nova-cell1-db-create-wjg5p" Oct 04 08:35:17 crc kubenswrapper[4969]: I1004 08:35:17.689593 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wjg5p" Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.378283 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.378633 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="ceilometer-central-agent" containerID="cri-o://b768e5ff863031298333e33ac095d23331b27d4256926c232249e752b9760ca0" gracePeriod=30 Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.379520 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="proxy-httpd" containerID="cri-o://25989098a2aa6cbda783ccefe89d7b2eb1ca44bb80e94c75353ebb8630868e1d" gracePeriod=30 Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.379573 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="sg-core" containerID="cri-o://4974cbb3139927f9e1a46aea06afe4cb091e883ecef931f80ed9502a336d98f0" gracePeriod=30 Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.379636 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="ceilometer-notification-agent" containerID="cri-o://6cc5b7b41370dedd9ec0b0a195bc591aa46868845bb4ecdda84c51c080a6e195" gracePeriod=30 Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.516947 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.960083 4969 generic.go:334] "Generic (PLEG): container finished" podID="56d85928-e2de-4594-8578-b504979d67fc" containerID="4974cbb3139927f9e1a46aea06afe4cb091e883ecef931f80ed9502a336d98f0" exitCode=2 Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.960166 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerDied","Data":"4974cbb3139927f9e1a46aea06afe4cb091e883ecef931f80ed9502a336d98f0"} Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.965997 4969 generic.go:334] "Generic (PLEG): container finished" podID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerID="1ea51f099c2095e8c4ad9a8a37772af38d046e5af3f1dd016991426b6c0807db" exitCode=137 Oct 04 08:35:18 crc kubenswrapper[4969]: I1004 08:35:18.966040 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586c975586-mxhpd" event={"ID":"99936cdc-7b05-421f-89e9-6e65ee6e2075","Type":"ContainerDied","Data":"1ea51f099c2095e8c4ad9a8a37772af38d046e5af3f1dd016991426b6c0807db"} Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.652708 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.669059 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.669131 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.735967 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b6dcf6db5-m8jj5"] Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.737369 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" podUID="d7d78c6c-c9e9-4a37-babc-801522168a84" containerName="dnsmasq-dns" containerID="cri-o://c6d8e4496044b41533943667c339f923c40c096a7ecc3cdb6a3dd26e0656626f" gracePeriod=10 Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.981286 4969 generic.go:334] "Generic (PLEG): container finished" podID="56d85928-e2de-4594-8578-b504979d67fc" containerID="25989098a2aa6cbda783ccefe89d7b2eb1ca44bb80e94c75353ebb8630868e1d" exitCode=0 Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.981316 4969 generic.go:334] "Generic (PLEG): container finished" podID="56d85928-e2de-4594-8578-b504979d67fc" containerID="b768e5ff863031298333e33ac095d23331b27d4256926c232249e752b9760ca0" exitCode=0 Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.981326 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerDied","Data":"25989098a2aa6cbda783ccefe89d7b2eb1ca44bb80e94c75353ebb8630868e1d"} Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.981370 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerDied","Data":"b768e5ff863031298333e33ac095d23331b27d4256926c232249e752b9760ca0"} Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.987041 4969 generic.go:334] "Generic (PLEG): container finished" podID="d7d78c6c-c9e9-4a37-babc-801522168a84" containerID="c6d8e4496044b41533943667c339f923c40c096a7ecc3cdb6a3dd26e0656626f" exitCode=0 Oct 04 08:35:19 crc kubenswrapper[4969]: I1004 08:35:19.987088 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" event={"ID":"d7d78c6c-c9e9-4a37-babc-801522168a84","Type":"ContainerDied","Data":"c6d8e4496044b41533943667c339f923c40c096a7ecc3cdb6a3dd26e0656626f"} Oct 04 08:35:21 crc kubenswrapper[4969]: I1004 08:35:21.007187 4969 generic.go:334] "Generic (PLEG): container finished" podID="56d85928-e2de-4594-8578-b504979d67fc" containerID="6cc5b7b41370dedd9ec0b0a195bc591aa46868845bb4ecdda84c51c080a6e195" exitCode=0 Oct 04 08:35:21 crc kubenswrapper[4969]: I1004 08:35:21.007266 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerDied","Data":"6cc5b7b41370dedd9ec0b0a195bc591aa46868845bb4ecdda84c51c080a6e195"} Oct 04 08:35:21 crc kubenswrapper[4969]: I1004 08:35:21.406541 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:21 crc kubenswrapper[4969]: I1004 08:35:21.412125 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:22 crc kubenswrapper[4969]: I1004 08:35:22.260160 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" podUID="d7d78c6c-c9e9-4a37-babc-801522168a84" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: connect: connection refused" Oct 04 08:35:22 crc kubenswrapper[4969]: I1004 08:35:22.817852 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:35:22 crc kubenswrapper[4969]: I1004 08:35:22.818194 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-569b746d9d-r7c45" Oct 04 08:35:24 crc kubenswrapper[4969]: I1004 08:35:24.366768 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:24 crc kubenswrapper[4969]: I1004 08:35:24.428792 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-586c975586-mxhpd" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Oct 04 08:35:24 crc kubenswrapper[4969]: I1004 08:35:24.439404 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dbff44956-qlvk2" Oct 04 08:35:24 crc kubenswrapper[4969]: I1004 08:35:24.492718 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-798dcfcb48-rr6vl"] Oct 04 08:35:24 crc kubenswrapper[4969]: I1004 08:35:24.494613 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-798dcfcb48-rr6vl" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api-log" containerID="cri-o://7e4ef3e6610aa4c53eca0a0bf956e1000c461ac0ca56693894ae8703622d3b98" gracePeriod=30 Oct 04 08:35:24 crc kubenswrapper[4969]: I1004 08:35:24.494727 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-798dcfcb48-rr6vl" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api" containerID="cri-o://569dd599a2462b73e18ddf2712ae36e6399c2b7b176b25bbd0a2393a5d7966cb" gracePeriod=30 Oct 04 08:35:24 crc kubenswrapper[4969]: I1004 08:35:24.625614 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 04 08:35:24 crc kubenswrapper[4969]: I1004 08:35:24.625691 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:35:24 crc kubenswrapper[4969]: I1004 08:35:24.626314 4969 scope.go:117] "RemoveContainer" containerID="a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.059215 4969 generic.go:334] "Generic (PLEG): container finished" podID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerID="7e4ef3e6610aa4c53eca0a0bf956e1000c461ac0ca56693894ae8703622d3b98" exitCode=143 Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.067468 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798dcfcb48-rr6vl" event={"ID":"67d580e7-abf5-476a-8aaa-09a8415193e3","Type":"ContainerDied","Data":"7e4ef3e6610aa4c53eca0a0bf956e1000c461ac0ca56693894ae8703622d3b98"} Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.366275 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-798dcfcb48-rr6vl" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.175:9311/healthcheck\": read tcp 10.217.0.2:47648->10.217.0.175:9311: read: connection reset by peer" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.366299 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-798dcfcb48-rr6vl" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.175:9311/healthcheck\": read tcp 10.217.0.2:47660->10.217.0.175:9311: read: connection reset by peer" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.570886 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.617152 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjfvn\" (UniqueName: \"kubernetes.io/projected/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-kube-api-access-bjfvn\") pod \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.617204 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-combined-ca-bundle\") pod \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.617244 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-config\") pod \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\" (UID: \"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.634055 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-kube-api-access-bjfvn" (OuterVolumeSpecName: "kube-api-access-bjfvn") pod "2861dd9f-5291-4f2a-bb30-0b9c54baa5cf" (UID: "2861dd9f-5291-4f2a-bb30-0b9c54baa5cf"). InnerVolumeSpecName "kube-api-access-bjfvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.681719 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2861dd9f-5291-4f2a-bb30-0b9c54baa5cf" (UID: "2861dd9f-5291-4f2a-bb30-0b9c54baa5cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.711617 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-config" (OuterVolumeSpecName: "config") pod "2861dd9f-5291-4f2a-bb30-0b9c54baa5cf" (UID: "2861dd9f-5291-4f2a-bb30-0b9c54baa5cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.719133 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.720501 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjfvn\" (UniqueName: \"kubernetes.io/projected/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-kube-api-access-bjfvn\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.720631 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.720710 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.824300 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc6zz\" (UniqueName: \"kubernetes.io/projected/99936cdc-7b05-421f-89e9-6e65ee6e2075-kube-api-access-zc6zz\") pod \"99936cdc-7b05-421f-89e9-6e65ee6e2075\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.824645 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-tls-certs\") pod \"99936cdc-7b05-421f-89e9-6e65ee6e2075\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.824809 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-config-data\") pod \"99936cdc-7b05-421f-89e9-6e65ee6e2075\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.824892 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-scripts\") pod \"99936cdc-7b05-421f-89e9-6e65ee6e2075\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.824995 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-combined-ca-bundle\") pod \"99936cdc-7b05-421f-89e9-6e65ee6e2075\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.825218 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99936cdc-7b05-421f-89e9-6e65ee6e2075-logs\") pod \"99936cdc-7b05-421f-89e9-6e65ee6e2075\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.825285 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-secret-key\") pod \"99936cdc-7b05-421f-89e9-6e65ee6e2075\" (UID: \"99936cdc-7b05-421f-89e9-6e65ee6e2075\") " Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.832779 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99936cdc-7b05-421f-89e9-6e65ee6e2075-logs" (OuterVolumeSpecName: "logs") pod "99936cdc-7b05-421f-89e9-6e65ee6e2075" (UID: "99936cdc-7b05-421f-89e9-6e65ee6e2075"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.833991 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "99936cdc-7b05-421f-89e9-6e65ee6e2075" (UID: "99936cdc-7b05-421f-89e9-6e65ee6e2075"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.840469 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99936cdc-7b05-421f-89e9-6e65ee6e2075-kube-api-access-zc6zz" (OuterVolumeSpecName: "kube-api-access-zc6zz") pod "99936cdc-7b05-421f-89e9-6e65ee6e2075" (UID: "99936cdc-7b05-421f-89e9-6e65ee6e2075"). InnerVolumeSpecName "kube-api-access-zc6zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.873452 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-scripts" (OuterVolumeSpecName: "scripts") pod "99936cdc-7b05-421f-89e9-6e65ee6e2075" (UID: "99936cdc-7b05-421f-89e9-6e65ee6e2075"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.925414 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-config-data" (OuterVolumeSpecName: "config-data") pod "99936cdc-7b05-421f-89e9-6e65ee6e2075" (UID: "99936cdc-7b05-421f-89e9-6e65ee6e2075"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.927315 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.927442 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99936cdc-7b05-421f-89e9-6e65ee6e2075-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.927501 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99936cdc-7b05-421f-89e9-6e65ee6e2075-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.927551 4969 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.927612 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc6zz\" (UniqueName: \"kubernetes.io/projected/99936cdc-7b05-421f-89e9-6e65ee6e2075-kube-api-access-zc6zz\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:25 crc kubenswrapper[4969]: I1004 08:35:25.971613 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.032463 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-combined-ca-bundle\") pod \"56d85928-e2de-4594-8578-b504979d67fc\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.032601 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-run-httpd\") pod \"56d85928-e2de-4594-8578-b504979d67fc\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.032659 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gl4p\" (UniqueName: \"kubernetes.io/projected/56d85928-e2de-4594-8578-b504979d67fc-kube-api-access-6gl4p\") pod \"56d85928-e2de-4594-8578-b504979d67fc\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.032678 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-scripts\") pod \"56d85928-e2de-4594-8578-b504979d67fc\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.032753 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-config-data\") pod \"56d85928-e2de-4594-8578-b504979d67fc\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.032779 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-sg-core-conf-yaml\") pod \"56d85928-e2de-4594-8578-b504979d67fc\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.032864 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-log-httpd\") pod \"56d85928-e2de-4594-8578-b504979d67fc\" (UID: \"56d85928-e2de-4594-8578-b504979d67fc\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.033623 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "56d85928-e2de-4594-8578-b504979d67fc" (UID: "56d85928-e2de-4594-8578-b504979d67fc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.047409 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.047787 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "56d85928-e2de-4594-8578-b504979d67fc" (UID: "56d85928-e2de-4594-8578-b504979d67fc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.056307 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-scripts" (OuterVolumeSpecName: "scripts") pod "56d85928-e2de-4594-8578-b504979d67fc" (UID: "56d85928-e2de-4594-8578-b504979d67fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.057344 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56d85928-e2de-4594-8578-b504979d67fc-kube-api-access-6gl4p" (OuterVolumeSpecName: "kube-api-access-6gl4p") pod "56d85928-e2de-4594-8578-b504979d67fc" (UID: "56d85928-e2de-4594-8578-b504979d67fc"). InnerVolumeSpecName "kube-api-access-6gl4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.066117 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99936cdc-7b05-421f-89e9-6e65ee6e2075" (UID: "99936cdc-7b05-421f-89e9-6e65ee6e2075"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.088534 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "99936cdc-7b05-421f-89e9-6e65ee6e2075" (UID: "99936cdc-7b05-421f-89e9-6e65ee6e2075"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.118194 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d85928-e2de-4594-8578-b504979d67fc","Type":"ContainerDied","Data":"e8b745d7981bf7bcf73e64eada07769197c859be75053d9131ee19c91c66db53"} Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.118542 4969 scope.go:117] "RemoveContainer" containerID="25989098a2aa6cbda783ccefe89d7b2eb1ca44bb80e94c75353ebb8630868e1d" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.118694 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.134228 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-svc\") pod \"d7d78c6c-c9e9-4a37-babc-801522168a84\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.134301 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-nb\") pod \"d7d78c6c-c9e9-4a37-babc-801522168a84\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.134556 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-swift-storage-0\") pod \"d7d78c6c-c9e9-4a37-babc-801522168a84\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.134670 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-sb\") pod \"d7d78c6c-c9e9-4a37-babc-801522168a84\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.134704 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cc76\" (UniqueName: \"kubernetes.io/projected/d7d78c6c-c9e9-4a37-babc-801522168a84-kube-api-access-5cc76\") pod \"d7d78c6c-c9e9-4a37-babc-801522168a84\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.134758 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-config\") pod \"d7d78c6c-c9e9-4a37-babc-801522168a84\" (UID: \"d7d78c6c-c9e9-4a37-babc-801522168a84\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.136468 4969 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.136492 4969 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.136502 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gl4p\" (UniqueName: \"kubernetes.io/projected/56d85928-e2de-4594-8578-b504979d67fc-kube-api-access-6gl4p\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.136513 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.136531 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99936cdc-7b05-421f-89e9-6e65ee6e2075-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.136539 4969 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d85928-e2de-4594-8578-b504979d67fc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.142811 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586c975586-mxhpd" event={"ID":"99936cdc-7b05-421f-89e9-6e65ee6e2075","Type":"ContainerDied","Data":"ad7176901b69399f44e9724125d3cd89d15fcd51666ca8eb8b47c46edfec6152"} Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.142984 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-586c975586-mxhpd" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.154193 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-7fphd"] Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.155165 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cf9ls" event={"ID":"2861dd9f-5291-4f2a-bb30-0b9c54baa5cf","Type":"ContainerDied","Data":"13ac15667bf398c16f59ba748239280cf910691d69de0ed27a6b8bbe5073bae2"} Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.155201 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13ac15667bf398c16f59ba748239280cf910691d69de0ed27a6b8bbe5073bae2" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.155268 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cf9ls" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.166023 4969 generic.go:334] "Generic (PLEG): container finished" podID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerID="569dd599a2462b73e18ddf2712ae36e6399c2b7b176b25bbd0a2393a5d7966cb" exitCode=0 Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.166099 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798dcfcb48-rr6vl" event={"ID":"67d580e7-abf5-476a-8aaa-09a8415193e3","Type":"ContainerDied","Data":"569dd599a2462b73e18ddf2712ae36e6399c2b7b176b25bbd0a2393a5d7966cb"} Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.169192 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-75f87ddd45-2md8q" event={"ID":"f2887852-0cd7-476e-8c4e-1ed98c66ede4","Type":"ContainerStarted","Data":"83256d50a6c876c9c00efe305b138bcdb5c5e29eefda0510dfe13c21de806259"} Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.177688 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" event={"ID":"d7d78c6c-c9e9-4a37-babc-801522168a84","Type":"ContainerDied","Data":"f52f38b3698b990789778ad1807ee76316451a626193cb0124c746c64450c32b"} Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.177760 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6dcf6db5-m8jj5" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.194620 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d78c6c-c9e9-4a37-babc-801522168a84-kube-api-access-5cc76" (OuterVolumeSpecName: "kube-api-access-5cc76") pod "d7d78c6c-c9e9-4a37-babc-801522168a84" (UID: "d7d78c6c-c9e9-4a37-babc-801522168a84"). InnerVolumeSpecName "kube-api-access-5cc76". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.225093 4969 scope.go:117] "RemoveContainer" containerID="4974cbb3139927f9e1a46aea06afe4cb091e883ecef931f80ed9502a336d98f0" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.234163 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "56d85928-e2de-4594-8578-b504979d67fc" (UID: "56d85928-e2de-4594-8578-b504979d67fc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.239298 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cc76\" (UniqueName: \"kubernetes.io/projected/d7d78c6c-c9e9-4a37-babc-801522168a84-kube-api-access-5cc76\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.239326 4969 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.393432 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d7d78c6c-c9e9-4a37-babc-801522168a84" (UID: "d7d78c6c-c9e9-4a37-babc-801522168a84"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.397814 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bx4tb"] Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.414030 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wjg5p"] Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.446670 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.490919 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-config" (OuterVolumeSpecName: "config") pod "d7d78c6c-c9e9-4a37-babc-801522168a84" (UID: "d7d78c6c-c9e9-4a37-babc-801522168a84"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.544387 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56d85928-e2de-4594-8578-b504979d67fc" (UID: "56d85928-e2de-4594-8578-b504979d67fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.548091 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.548124 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.560863 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7d78c6c-c9e9-4a37-babc-801522168a84" (UID: "d7d78c6c-c9e9-4a37-babc-801522168a84"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.570837 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7d78c6c-c9e9-4a37-babc-801522168a84" (UID: "d7d78c6c-c9e9-4a37-babc-801522168a84"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.575047 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-config-data" (OuterVolumeSpecName: "config-data") pod "56d85928-e2de-4594-8578-b504979d67fc" (UID: "56d85928-e2de-4594-8578-b504979d67fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.575573 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7d78c6c-c9e9-4a37-babc-801522168a84" (UID: "d7d78c6c-c9e9-4a37-babc-801522168a84"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.650375 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.650522 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d85928-e2de-4594-8578-b504979d67fc-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.650539 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.650555 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d78c6c-c9e9-4a37-babc-801522168a84-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.673592 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.680469 4969 scope.go:117] "RemoveContainer" containerID="6cc5b7b41370dedd9ec0b0a195bc591aa46868845bb4ecdda84c51c080a6e195" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.693590 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-586c975586-mxhpd"] Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.712488 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-586c975586-mxhpd"] Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.751883 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data-custom\") pod \"67d580e7-abf5-476a-8aaa-09a8415193e3\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.751953 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-combined-ca-bundle\") pod \"67d580e7-abf5-476a-8aaa-09a8415193e3\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.751985 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d580e7-abf5-476a-8aaa-09a8415193e3-logs\") pod \"67d580e7-abf5-476a-8aaa-09a8415193e3\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.752055 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data\") pod \"67d580e7-abf5-476a-8aaa-09a8415193e3\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.752080 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5hzl\" (UniqueName: \"kubernetes.io/projected/67d580e7-abf5-476a-8aaa-09a8415193e3-kube-api-access-b5hzl\") pod \"67d580e7-abf5-476a-8aaa-09a8415193e3\" (UID: \"67d580e7-abf5-476a-8aaa-09a8415193e3\") " Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.754878 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d580e7-abf5-476a-8aaa-09a8415193e3-logs" (OuterVolumeSpecName: "logs") pod "67d580e7-abf5-476a-8aaa-09a8415193e3" (UID: "67d580e7-abf5-476a-8aaa-09a8415193e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.777415 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d580e7-abf5-476a-8aaa-09a8415193e3-kube-api-access-b5hzl" (OuterVolumeSpecName: "kube-api-access-b5hzl") pod "67d580e7-abf5-476a-8aaa-09a8415193e3" (UID: "67d580e7-abf5-476a-8aaa-09a8415193e3"). InnerVolumeSpecName "kube-api-access-b5hzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.786482 4969 scope.go:117] "RemoveContainer" containerID="b768e5ff863031298333e33ac095d23331b27d4256926c232249e752b9760ca0" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.818895 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "67d580e7-abf5-476a-8aaa-09a8415193e3" (UID: "67d580e7-abf5-476a-8aaa-09a8415193e3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.855655 4969 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.855686 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d580e7-abf5-476a-8aaa-09a8415193e3-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.855695 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5hzl\" (UniqueName: \"kubernetes.io/projected/67d580e7-abf5-476a-8aaa-09a8415193e3-kube-api-access-b5hzl\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.865979 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b8d76ff-7kf7d"] Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866410 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866440 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866451 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d78c6c-c9e9-4a37-babc-801522168a84" containerName="init" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866461 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d78c6c-c9e9-4a37-babc-801522168a84" containerName="init" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866477 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2861dd9f-5291-4f2a-bb30-0b9c54baa5cf" containerName="neutron-db-sync" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866486 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="2861dd9f-5291-4f2a-bb30-0b9c54baa5cf" containerName="neutron-db-sync" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866499 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="ceilometer-notification-agent" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866507 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="ceilometer-notification-agent" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866517 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866525 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866550 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon-log" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866557 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon-log" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866576 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d78c6c-c9e9-4a37-babc-801522168a84" containerName="dnsmasq-dns" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866581 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d78c6c-c9e9-4a37-babc-801522168a84" containerName="dnsmasq-dns" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866593 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="ceilometer-central-agent" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866599 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="ceilometer-central-agent" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866617 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="proxy-httpd" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866623 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="proxy-httpd" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866634 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api-log" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866641 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api-log" Oct 04 08:35:26 crc kubenswrapper[4969]: E1004 08:35:26.866651 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="sg-core" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866658 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="sg-core" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866836 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="ceilometer-notification-agent" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866848 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="ceilometer-central-agent" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866857 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="proxy-httpd" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866867 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d78c6c-c9e9-4a37-babc-801522168a84" containerName="dnsmasq-dns" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866880 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866893 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" containerName="barbican-api-log" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866909 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d85928-e2de-4594-8578-b504979d67fc" containerName="sg-core" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866922 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866927 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="2861dd9f-5291-4f2a-bb30-0b9c54baa5cf" containerName="neutron-db-sync" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.866941 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" containerName="horizon-log" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.867943 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.880109 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b8d76ff-7kf7d"] Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.957949 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-sb\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.957986 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-svc\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.958023 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-nb\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.958084 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-swift-storage-0\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.958139 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8np22\" (UniqueName: \"kubernetes.io/projected/6a27a749-7256-45fc-b79c-4882d5554e9d-kube-api-access-8np22\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.958156 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-config\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:26 crc kubenswrapper[4969]: I1004 08:35:26.961469 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67d580e7-abf5-476a-8aaa-09a8415193e3" (UID: "67d580e7-abf5-476a-8aaa-09a8415193e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.052310 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data" (OuterVolumeSpecName: "config-data") pod "67d580e7-abf5-476a-8aaa-09a8415193e3" (UID: "67d580e7-abf5-476a-8aaa-09a8415193e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.066405 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8np22\" (UniqueName: \"kubernetes.io/projected/6a27a749-7256-45fc-b79c-4882d5554e9d-kube-api-access-8np22\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.066452 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-config\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.066498 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-sb\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.066515 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-svc\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.066546 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-nb\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.066602 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-swift-storage-0\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.066664 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.066675 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d580e7-abf5-476a-8aaa-09a8415193e3-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.067353 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-swift-storage-0\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.070546 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-sb\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.071101 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-svc\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.071952 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-nb\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.072404 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-config\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.075830 4969 scope.go:117] "RemoveContainer" containerID="d7a1460398fc911e418d1988d141295bff2e22e358f4021fa1bf97a2960e568e" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.098204 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99936cdc-7b05-421f-89e9-6e65ee6e2075" path="/var/lib/kubelet/pods/99936cdc-7b05-421f-89e9-6e65ee6e2075/volumes" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.100099 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-86d475f9d4-dkrh5"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.102000 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b6dcf6db5-m8jj5"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.102087 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.106269 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.106487 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ct9bq" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.106994 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.109107 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.111849 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8np22\" (UniqueName: \"kubernetes.io/projected/6a27a749-7256-45fc-b79c-4882d5554e9d-kube-api-access-8np22\") pod \"dnsmasq-dns-666b8d76ff-7kf7d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.159631 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b6dcf6db5-m8jj5"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.167896 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-ovndb-tls-certs\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.168214 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-combined-ca-bundle\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.168319 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-config\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.168510 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkjdp\" (UniqueName: \"kubernetes.io/projected/47ecd91e-5401-418c-a830-6a67c0965cc4-kube-api-access-kkjdp\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.168779 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-httpd-config\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.174530 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86d475f9d4-dkrh5"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.184166 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.192486 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.194431 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bx4tb" event={"ID":"f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f","Type":"ContainerStarted","Data":"65ea1e18cdd39af7f4a6148bf32f5950ef0fd483cbc159d23dca96f64aecff13"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.194486 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bx4tb" event={"ID":"f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f","Type":"ContainerStarted","Data":"4a9ef8a1df533046d09b7ea6a29a25e02d98bd049122348e8bb379dcda3a34cb"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.220291 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798dcfcb48-rr6vl" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.220271 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798dcfcb48-rr6vl" event={"ID":"67d580e7-abf5-476a-8aaa-09a8415193e3","Type":"ContainerDied","Data":"9d9da6d45f95742828c1c159963971908f4992faa1215e58c25729f7679aa132"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.226973 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7fphd" event={"ID":"1ff3f16f-8103-49a7-8a5c-07b6b7f399e7","Type":"ContainerStarted","Data":"51525ee67e0a0f9a69dc004390b4adfb3a6e5bdcf597d1a88687b00712ab9cc7"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.227024 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7fphd" event={"ID":"1ff3f16f-8103-49a7-8a5c-07b6b7f399e7","Type":"ContainerStarted","Data":"96cdeefe3cbebcf072b0b7d4316dd2b09bff836fb8e57b203ca1516cc449b637"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.228399 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wjg5p" event={"ID":"46a217c4-ea44-4608-9300-65410bec2b88","Type":"ContainerStarted","Data":"b0104d6358c5bb555eb7989ed09b665985b991cfd2c024072e2041c22a65a774"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.234624 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f5532f51-e305-4741-8661-25029faf98f2","Type":"ContainerStarted","Data":"f84d231cfc5dd7220af66da52d9e8e0ae755fb864b1a5491a29bc0db1ee98b04"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.236205 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.245219 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.248581 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.248843 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.257262 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-75f87ddd45-2md8q" event={"ID":"f2887852-0cd7-476e-8c4e-1ed98c66ede4","Type":"ContainerStarted","Data":"cf3f70cad6b43af1d0cc61310adb158452d7b2a6f7750ed5180083d090532f83"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.257470 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-75f87ddd45-2md8q" event={"ID":"f2887852-0cd7-476e-8c4e-1ed98c66ede4","Type":"ContainerStarted","Data":"6e1ae0fee4bee96d5190e2f3ef0ddce6a4abe498676bdbb55a10164491f55390"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.257702 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.257958 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.260159 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerStarted","Data":"84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58"} Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.271910 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-run-httpd\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.271959 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-ovndb-tls-certs\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.271991 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-scripts\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.272024 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-combined-ca-bundle\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.272051 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.272076 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-config\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.272113 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkjdp\" (UniqueName: \"kubernetes.io/projected/47ecd91e-5401-418c-a830-6a67c0965cc4-kube-api-access-kkjdp\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.272165 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.272200 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-config-data\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.272293 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z28sn\" (UniqueName: \"kubernetes.io/projected/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-kube-api-access-z28sn\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.272320 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-log-httpd\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.272344 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-httpd-config\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.281740 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-httpd-config\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.283617 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-config\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.289484 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.298955 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-ovndb-tls-certs\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.299936 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkjdp\" (UniqueName: \"kubernetes.io/projected/47ecd91e-5401-418c-a830-6a67c0965cc4-kube-api-access-kkjdp\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.307678 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-combined-ca-bundle\") pod \"neutron-86d475f9d4-dkrh5\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.325743 4969 scope.go:117] "RemoveContainer" containerID="1ea51f099c2095e8c4ad9a8a37772af38d046e5af3f1dd016991426b6c0807db" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.363079 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-bx4tb" podStartSLOduration=10.363058139 podStartE2EDuration="10.363058139s" podCreationTimestamp="2025-10-04 08:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:27.212880293 +0000 UTC m=+1154.967149107" watchObservedRunningTime="2025-10-04 08:35:27.363058139 +0000 UTC m=+1155.117326953" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.368281 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.373895 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-run-httpd\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.373926 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-scripts\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.373960 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.374022 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.374054 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-config-data\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.374120 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z28sn\" (UniqueName: \"kubernetes.io/projected/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-kube-api-access-z28sn\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.374138 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-log-httpd\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.383084 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-run-httpd\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.384844 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-7fphd" podStartSLOduration=10.384823181 podStartE2EDuration="10.384823181s" podCreationTimestamp="2025-10-04 08:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:27.263112919 +0000 UTC m=+1155.017381733" watchObservedRunningTime="2025-10-04 08:35:27.384823181 +0000 UTC m=+1155.139091995" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.392933 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.299938806 podStartE2EDuration="16.392916888s" podCreationTimestamp="2025-10-04 08:35:11 +0000 UTC" firstStartedPulling="2025-10-04 08:35:13.383547152 +0000 UTC m=+1141.137815966" lastFinishedPulling="2025-10-04 08:35:25.476525234 +0000 UTC m=+1153.230794048" observedRunningTime="2025-10-04 08:35:27.276539887 +0000 UTC m=+1155.030808711" watchObservedRunningTime="2025-10-04 08:35:27.392916888 +0000 UTC m=+1155.147185702" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.398405 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-log-httpd\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.398829 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.401080 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-config-data\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.402358 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-scripts\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.409916 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.413829 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z28sn\" (UniqueName: \"kubernetes.io/projected/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-kube-api-access-z28sn\") pod \"ceilometer-0\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.423602 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-798dcfcb48-rr6vl"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.434629 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.435271 4969 scope.go:117] "RemoveContainer" containerID="c6d8e4496044b41533943667c339f923c40c096a7ecc3cdb6a3dd26e0656626f" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.445514 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-798dcfcb48-rr6vl"] Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.450365 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-75f87ddd45-2md8q" podStartSLOduration=11.45034833 podStartE2EDuration="11.45034833s" podCreationTimestamp="2025-10-04 08:35:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:27.358772964 +0000 UTC m=+1155.113041778" watchObservedRunningTime="2025-10-04 08:35:27.45034833 +0000 UTC m=+1155.204617144" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.502634 4969 scope.go:117] "RemoveContainer" containerID="9b070738e94db6fa89d7792e706587342acd0ab263810ad74ec784aca57447b0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.568928 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.581279 4969 scope.go:117] "RemoveContainer" containerID="569dd599a2462b73e18ddf2712ae36e6399c2b7b176b25bbd0a2393a5d7966cb" Oct 04 08:35:27 crc kubenswrapper[4969]: I1004 08:35:27.665502 4969 scope.go:117] "RemoveContainer" containerID="7e4ef3e6610aa4c53eca0a0bf956e1000c461ac0ca56693894ae8703622d3b98" Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.085370 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86d475f9d4-dkrh5"] Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.106663 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b8d76ff-7kf7d"] Oct 04 08:35:28 crc kubenswrapper[4969]: W1004 08:35:28.109745 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a27a749_7256_45fc_b79c_4882d5554e9d.slice/crio-2145e3b8b65ed6b486e1c75f271e3ac34d63e3b86a3a9f4533b9cb50ebc5ca9c WatchSource:0}: Error finding container 2145e3b8b65ed6b486e1c75f271e3ac34d63e3b86a3a9f4533b9cb50ebc5ca9c: Status 404 returned error can't find the container with id 2145e3b8b65ed6b486e1c75f271e3ac34d63e3b86a3a9f4533b9cb50ebc5ca9c Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.268143 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.280753 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" event={"ID":"6a27a749-7256-45fc-b79c-4882d5554e9d","Type":"ContainerStarted","Data":"2145e3b8b65ed6b486e1c75f271e3ac34d63e3b86a3a9f4533b9cb50ebc5ca9c"} Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.286547 4969 generic.go:334] "Generic (PLEG): container finished" podID="1ff3f16f-8103-49a7-8a5c-07b6b7f399e7" containerID="51525ee67e0a0f9a69dc004390b4adfb3a6e5bdcf597d1a88687b00712ab9cc7" exitCode=0 Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.286619 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7fphd" event={"ID":"1ff3f16f-8103-49a7-8a5c-07b6b7f399e7","Type":"ContainerDied","Data":"51525ee67e0a0f9a69dc004390b4adfb3a6e5bdcf597d1a88687b00712ab9cc7"} Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.303676 4969 generic.go:334] "Generic (PLEG): container finished" podID="46a217c4-ea44-4608-9300-65410bec2b88" containerID="ca6768d0d1b8429421ce74128dc645f9edaaf7b04b4952878c75bdd98d883fc3" exitCode=0 Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.303800 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wjg5p" event={"ID":"46a217c4-ea44-4608-9300-65410bec2b88","Type":"ContainerDied","Data":"ca6768d0d1b8429421ce74128dc645f9edaaf7b04b4952878c75bdd98d883fc3"} Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.318995 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86d475f9d4-dkrh5" event={"ID":"47ecd91e-5401-418c-a830-6a67c0965cc4","Type":"ContainerStarted","Data":"c96daad6e33b4ff7512c225598849614639fe61d5de22debd74e31b95208531d"} Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.329978 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xjc7p" event={"ID":"278ebf83-6c66-4d97-8b5c-a7a83ace00b7","Type":"ContainerStarted","Data":"449fe9b996875ca6746e78842bfb0e9eefbf4e3ad978360a06766e20c54204a8"} Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.334462 4969 generic.go:334] "Generic (PLEG): container finished" podID="f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f" containerID="65ea1e18cdd39af7f4a6148bf32f5950ef0fd483cbc159d23dca96f64aecff13" exitCode=0 Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.334519 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bx4tb" event={"ID":"f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f","Type":"ContainerDied","Data":"65ea1e18cdd39af7f4a6148bf32f5950ef0fd483cbc159d23dca96f64aecff13"} Oct 04 08:35:28 crc kubenswrapper[4969]: I1004 08:35:28.350666 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-xjc7p" podStartSLOduration=6.609164253 podStartE2EDuration="57.350648888s" podCreationTimestamp="2025-10-04 08:34:31 +0000 UTC" firstStartedPulling="2025-10-04 08:34:34.949112105 +0000 UTC m=+1102.703380919" lastFinishedPulling="2025-10-04 08:35:25.69059674 +0000 UTC m=+1153.444865554" observedRunningTime="2025-10-04 08:35:28.349728527 +0000 UTC m=+1156.103997341" watchObservedRunningTime="2025-10-04 08:35:28.350648888 +0000 UTC m=+1156.104917702" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.065880 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56d85928-e2de-4594-8578-b504979d67fc" path="/var/lib/kubelet/pods/56d85928-e2de-4594-8578-b504979d67fc/volumes" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.066852 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d580e7-abf5-476a-8aaa-09a8415193e3" path="/var/lib/kubelet/pods/67d580e7-abf5-476a-8aaa-09a8415193e3/volumes" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.067366 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d78c6c-c9e9-4a37-babc-801522168a84" path="/var/lib/kubelet/pods/d7d78c6c-c9e9-4a37-babc-801522168a84/volumes" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.351766 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86d475f9d4-dkrh5" event={"ID":"47ecd91e-5401-418c-a830-6a67c0965cc4","Type":"ContainerStarted","Data":"b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4"} Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.351810 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86d475f9d4-dkrh5" event={"ID":"47ecd91e-5401-418c-a830-6a67c0965cc4","Type":"ContainerStarted","Data":"729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915"} Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.352027 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.362649 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.366728 4969 generic.go:334] "Generic (PLEG): container finished" podID="6a27a749-7256-45fc-b79c-4882d5554e9d" containerID="4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056" exitCode=0 Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.366836 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" event={"ID":"6a27a749-7256-45fc-b79c-4882d5554e9d","Type":"ContainerDied","Data":"4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056"} Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.371391 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-86d475f9d4-dkrh5" podStartSLOduration=2.371374057 podStartE2EDuration="2.371374057s" podCreationTimestamp="2025-10-04 08:35:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:29.371326826 +0000 UTC m=+1157.125595640" watchObservedRunningTime="2025-10-04 08:35:29.371374057 +0000 UTC m=+1157.125642871" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.392634 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerStarted","Data":"f3bea43bcb525e61d96f68baeb53976065f2228a56da56ca315fda00d39d8072"} Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.392672 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerStarted","Data":"2fe691cbfcceb510d03848ec51e76ed42dd3d4cdbd6f5c652bf7a3d41fa5bc1f"} Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.392683 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerStarted","Data":"78ae6c6bfa8f82189420a489c90836f4adba999ebc3beb7bca4fbb3d66caae20"} Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.459625 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-78c8455845-gfrwq"] Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.461604 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.473140 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.473208 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.521555 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-ovndb-tls-certs\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.521686 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-httpd-config\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.521738 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-config\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.521796 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-internal-tls-certs\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.521874 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc287\" (UniqueName: \"kubernetes.io/projected/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-kube-api-access-cc287\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.521976 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-public-tls-certs\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.522146 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-combined-ca-bundle\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.561529 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78c8455845-gfrwq"] Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.628201 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-combined-ca-bundle\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.628276 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-ovndb-tls-certs\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.628307 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-httpd-config\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.628330 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-config\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.628357 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-internal-tls-certs\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.628384 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc287\" (UniqueName: \"kubernetes.io/projected/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-kube-api-access-cc287\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.628431 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-public-tls-certs\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.632801 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-public-tls-certs\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.632976 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-httpd-config\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.636111 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-internal-tls-certs\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.636484 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-ovndb-tls-certs\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.639782 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-config\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.646387 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-combined-ca-bundle\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.661203 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc287\" (UniqueName: \"kubernetes.io/projected/e82aaa27-6e5c-43d8-9ed2-c0958de485c7-kube-api-access-cc287\") pod \"neutron-78c8455845-gfrwq\" (UID: \"e82aaa27-6e5c-43d8-9ed2-c0958de485c7\") " pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.859921 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:29 crc kubenswrapper[4969]: I1004 08:35:29.954127 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bx4tb" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.040769 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s54n\" (UniqueName: \"kubernetes.io/projected/f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f-kube-api-access-4s54n\") pod \"f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f\" (UID: \"f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f\") " Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.049867 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f-kube-api-access-4s54n" (OuterVolumeSpecName: "kube-api-access-4s54n") pod "f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f" (UID: "f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f"). InnerVolumeSpecName "kube-api-access-4s54n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.079758 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wjg5p" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.115636 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7fphd" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.143664 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s54n\" (UniqueName: \"kubernetes.io/projected/f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f-kube-api-access-4s54n\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.244379 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfzlj\" (UniqueName: \"kubernetes.io/projected/1ff3f16f-8103-49a7-8a5c-07b6b7f399e7-kube-api-access-pfzlj\") pod \"1ff3f16f-8103-49a7-8a5c-07b6b7f399e7\" (UID: \"1ff3f16f-8103-49a7-8a5c-07b6b7f399e7\") " Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.244476 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnp9p\" (UniqueName: \"kubernetes.io/projected/46a217c4-ea44-4608-9300-65410bec2b88-kube-api-access-dnp9p\") pod \"46a217c4-ea44-4608-9300-65410bec2b88\" (UID: \"46a217c4-ea44-4608-9300-65410bec2b88\") " Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.250379 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ff3f16f-8103-49a7-8a5c-07b6b7f399e7-kube-api-access-pfzlj" (OuterVolumeSpecName: "kube-api-access-pfzlj") pod "1ff3f16f-8103-49a7-8a5c-07b6b7f399e7" (UID: "1ff3f16f-8103-49a7-8a5c-07b6b7f399e7"). InnerVolumeSpecName "kube-api-access-pfzlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.253040 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46a217c4-ea44-4608-9300-65410bec2b88-kube-api-access-dnp9p" (OuterVolumeSpecName: "kube-api-access-dnp9p") pod "46a217c4-ea44-4608-9300-65410bec2b88" (UID: "46a217c4-ea44-4608-9300-65410bec2b88"). InnerVolumeSpecName "kube-api-access-dnp9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.349544 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfzlj\" (UniqueName: \"kubernetes.io/projected/1ff3f16f-8103-49a7-8a5c-07b6b7f399e7-kube-api-access-pfzlj\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.349790 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnp9p\" (UniqueName: \"kubernetes.io/projected/46a217c4-ea44-4608-9300-65410bec2b88-kube-api-access-dnp9p\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.426918 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerStarted","Data":"f669ce1e61928245136e60acac4068532ec6ad12da11e683b5d1cc880b6ea8a4"} Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.444682 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wjg5p" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.444669 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wjg5p" event={"ID":"46a217c4-ea44-4608-9300-65410bec2b88","Type":"ContainerDied","Data":"b0104d6358c5bb555eb7989ed09b665985b991cfd2c024072e2041c22a65a774"} Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.444770 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0104d6358c5bb555eb7989ed09b665985b991cfd2c024072e2041c22a65a774" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.446554 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" event={"ID":"6a27a749-7256-45fc-b79c-4882d5554e9d","Type":"ContainerStarted","Data":"ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b"} Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.447652 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.457028 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bx4tb" event={"ID":"f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f","Type":"ContainerDied","Data":"4a9ef8a1df533046d09b7ea6a29a25e02d98bd049122348e8bb379dcda3a34cb"} Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.457067 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a9ef8a1df533046d09b7ea6a29a25e02d98bd049122348e8bb379dcda3a34cb" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.457120 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bx4tb" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.465511 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7fphd" event={"ID":"1ff3f16f-8103-49a7-8a5c-07b6b7f399e7","Type":"ContainerDied","Data":"96cdeefe3cbebcf072b0b7d4316dd2b09bff836fb8e57b203ca1516cc449b637"} Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.465546 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96cdeefe3cbebcf072b0b7d4316dd2b09bff836fb8e57b203ca1516cc449b637" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.465777 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7fphd" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.486111 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" podStartSLOduration=4.486083561 podStartE2EDuration="4.486083561s" podCreationTimestamp="2025-10-04 08:35:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:30.476044385 +0000 UTC m=+1158.230313219" watchObservedRunningTime="2025-10-04 08:35:30.486083561 +0000 UTC m=+1158.240352375" Oct 04 08:35:30 crc kubenswrapper[4969]: I1004 08:35:30.550389 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78c8455845-gfrwq"] Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.445557 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.497055 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerStarted","Data":"ee6efd9bdc504b0f26843761f74cd6eec1277fc36bc5b8b2bf80078e681a7639"} Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.498353 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="ceilometer-central-agent" containerID="cri-o://2fe691cbfcceb510d03848ec51e76ed42dd3d4cdbd6f5c652bf7a3d41fa5bc1f" gracePeriod=30 Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.498863 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="sg-core" containerID="cri-o://f669ce1e61928245136e60acac4068532ec6ad12da11e683b5d1cc880b6ea8a4" gracePeriod=30 Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.498966 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="proxy-httpd" containerID="cri-o://ee6efd9bdc504b0f26843761f74cd6eec1277fc36bc5b8b2bf80078e681a7639" gracePeriod=30 Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.499010 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="ceilometer-notification-agent" containerID="cri-o://f3bea43bcb525e61d96f68baeb53976065f2228a56da56ca315fda00d39d8072" gracePeriod=30 Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.501129 4969 generic.go:334] "Generic (PLEG): container finished" podID="7144eb34-59d2-4583-8c46-3e56258966da" containerID="84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58" exitCode=1 Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.501235 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerDied","Data":"84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58"} Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.501282 4969 scope.go:117] "RemoveContainer" containerID="a2ffbb32f22c9071b8ffc7ee1af9d04ff75068756ec1005852960eb54f2a29d6" Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.502165 4969 scope.go:117] "RemoveContainer" containerID="84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58" Oct 04 08:35:31 crc kubenswrapper[4969]: E1004 08:35:31.502650 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.510537 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78c8455845-gfrwq" event={"ID":"e82aaa27-6e5c-43d8-9ed2-c0958de485c7","Type":"ContainerStarted","Data":"2e6d123ac44299b8e6a70f6144d0a42c20361a3d23b214b475f549b0fa4188e3"} Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.510599 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78c8455845-gfrwq" event={"ID":"e82aaa27-6e5c-43d8-9ed2-c0958de485c7","Type":"ContainerStarted","Data":"3a9776f0a44ceee51060a4caef7df951730be7f3328f2da6ffd38ca0329258d3"} Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.510614 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78c8455845-gfrwq" event={"ID":"e82aaa27-6e5c-43d8-9ed2-c0958de485c7","Type":"ContainerStarted","Data":"541764406f4c1e876bfa8c658330d6ce47fc7593ff69c8fd2674970c833bf961"} Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.511761 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.544155 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.059807662 podStartE2EDuration="4.54413629s" podCreationTimestamp="2025-10-04 08:35:27 +0000 UTC" firstStartedPulling="2025-10-04 08:35:28.313143854 +0000 UTC m=+1156.067412658" lastFinishedPulling="2025-10-04 08:35:30.797472472 +0000 UTC m=+1158.551741286" observedRunningTime="2025-10-04 08:35:31.525343232 +0000 UTC m=+1159.279612056" watchObservedRunningTime="2025-10-04 08:35:31.54413629 +0000 UTC m=+1159.298405104" Oct 04 08:35:31 crc kubenswrapper[4969]: I1004 08:35:31.553222 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-78c8455845-gfrwq" podStartSLOduration=2.553203292 podStartE2EDuration="2.553203292s" podCreationTimestamp="2025-10-04 08:35:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:31.547643966 +0000 UTC m=+1159.301912790" watchObservedRunningTime="2025-10-04 08:35:31.553203292 +0000 UTC m=+1159.307472106" Oct 04 08:35:32 crc kubenswrapper[4969]: I1004 08:35:32.527327 4969 generic.go:334] "Generic (PLEG): container finished" podID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerID="ee6efd9bdc504b0f26843761f74cd6eec1277fc36bc5b8b2bf80078e681a7639" exitCode=0 Oct 04 08:35:32 crc kubenswrapper[4969]: I1004 08:35:32.528205 4969 generic.go:334] "Generic (PLEG): container finished" podID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerID="f669ce1e61928245136e60acac4068532ec6ad12da11e683b5d1cc880b6ea8a4" exitCode=2 Oct 04 08:35:32 crc kubenswrapper[4969]: I1004 08:35:32.528301 4969 generic.go:334] "Generic (PLEG): container finished" podID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerID="f3bea43bcb525e61d96f68baeb53976065f2228a56da56ca315fda00d39d8072" exitCode=0 Oct 04 08:35:32 crc kubenswrapper[4969]: I1004 08:35:32.527406 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerDied","Data":"ee6efd9bdc504b0f26843761f74cd6eec1277fc36bc5b8b2bf80078e681a7639"} Oct 04 08:35:32 crc kubenswrapper[4969]: I1004 08:35:32.528544 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerDied","Data":"f669ce1e61928245136e60acac4068532ec6ad12da11e683b5d1cc880b6ea8a4"} Oct 04 08:35:32 crc kubenswrapper[4969]: I1004 08:35:32.528653 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerDied","Data":"f3bea43bcb525e61d96f68baeb53976065f2228a56da56ca315fda00d39d8072"} Oct 04 08:35:34 crc kubenswrapper[4969]: I1004 08:35:34.551208 4969 generic.go:334] "Generic (PLEG): container finished" podID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerID="2fe691cbfcceb510d03848ec51e76ed42dd3d4cdbd6f5c652bf7a3d41fa5bc1f" exitCode=0 Oct 04 08:35:34 crc kubenswrapper[4969]: I1004 08:35:34.551508 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerDied","Data":"2fe691cbfcceb510d03848ec51e76ed42dd3d4cdbd6f5c652bf7a3d41fa5bc1f"} Oct 04 08:35:34 crc kubenswrapper[4969]: I1004 08:35:34.622373 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:35:34 crc kubenswrapper[4969]: I1004 08:35:34.622686 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:35:34 crc kubenswrapper[4969]: I1004 08:35:34.623381 4969 scope.go:117] "RemoveContainer" containerID="84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58" Oct 04 08:35:34 crc kubenswrapper[4969]: E1004 08:35:34.623665 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:35:34 crc kubenswrapper[4969]: I1004 08:35:34.852885 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.049866 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-log-httpd\") pod \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.050273 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-combined-ca-bundle\") pod \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.050314 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-sg-core-conf-yaml\") pod \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.050361 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-config-data\") pod \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.050348 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "309264c4-c0cc-4ec7-9f90-6aed3a3379ce" (UID: "309264c4-c0cc-4ec7-9f90-6aed3a3379ce"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.050514 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-run-httpd\") pod \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.050540 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-scripts\") pod \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.050754 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "309264c4-c0cc-4ec7-9f90-6aed3a3379ce" (UID: "309264c4-c0cc-4ec7-9f90-6aed3a3379ce"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.050808 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z28sn\" (UniqueName: \"kubernetes.io/projected/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-kube-api-access-z28sn\") pod \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\" (UID: \"309264c4-c0cc-4ec7-9f90-6aed3a3379ce\") " Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.051442 4969 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.051678 4969 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.058461 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-kube-api-access-z28sn" (OuterVolumeSpecName: "kube-api-access-z28sn") pod "309264c4-c0cc-4ec7-9f90-6aed3a3379ce" (UID: "309264c4-c0cc-4ec7-9f90-6aed3a3379ce"). InnerVolumeSpecName "kube-api-access-z28sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.059625 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-scripts" (OuterVolumeSpecName: "scripts") pod "309264c4-c0cc-4ec7-9f90-6aed3a3379ce" (UID: "309264c4-c0cc-4ec7-9f90-6aed3a3379ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.093559 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "309264c4-c0cc-4ec7-9f90-6aed3a3379ce" (UID: "309264c4-c0cc-4ec7-9f90-6aed3a3379ce"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.152507 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.152695 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z28sn\" (UniqueName: \"kubernetes.io/projected/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-kube-api-access-z28sn\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.152777 4969 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.170664 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "309264c4-c0cc-4ec7-9f90-6aed3a3379ce" (UID: "309264c4-c0cc-4ec7-9f90-6aed3a3379ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.220120 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-config-data" (OuterVolumeSpecName: "config-data") pod "309264c4-c0cc-4ec7-9f90-6aed3a3379ce" (UID: "309264c4-c0cc-4ec7-9f90-6aed3a3379ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.253436 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.253468 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309264c4-c0cc-4ec7-9f90-6aed3a3379ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.562544 4969 generic.go:334] "Generic (PLEG): container finished" podID="278ebf83-6c66-4d97-8b5c-a7a83ace00b7" containerID="449fe9b996875ca6746e78842bfb0e9eefbf4e3ad978360a06766e20c54204a8" exitCode=0 Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.562635 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xjc7p" event={"ID":"278ebf83-6c66-4d97-8b5c-a7a83ace00b7","Type":"ContainerDied","Data":"449fe9b996875ca6746e78842bfb0e9eefbf4e3ad978360a06766e20c54204a8"} Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.566027 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"309264c4-c0cc-4ec7-9f90-6aed3a3379ce","Type":"ContainerDied","Data":"78ae6c6bfa8f82189420a489c90836f4adba999ebc3beb7bca4fbb3d66caae20"} Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.566083 4969 scope.go:117] "RemoveContainer" containerID="ee6efd9bdc504b0f26843761f74cd6eec1277fc36bc5b8b2bf80078e681a7639" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.566094 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.599981 4969 scope.go:117] "RemoveContainer" containerID="f669ce1e61928245136e60acac4068532ec6ad12da11e683b5d1cc880b6ea8a4" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.644146 4969 scope.go:117] "RemoveContainer" containerID="f3bea43bcb525e61d96f68baeb53976065f2228a56da56ca315fda00d39d8072" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.653231 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.686339 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.694799 4969 scope.go:117] "RemoveContainer" containerID="2fe691cbfcceb510d03848ec51e76ed42dd3d4cdbd6f5c652bf7a3d41fa5bc1f" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716190 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:35 crc kubenswrapper[4969]: E1004 08:35:35.716649 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a217c4-ea44-4608-9300-65410bec2b88" containerName="mariadb-database-create" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716661 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a217c4-ea44-4608-9300-65410bec2b88" containerName="mariadb-database-create" Oct 04 08:35:35 crc kubenswrapper[4969]: E1004 08:35:35.716677 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="ceilometer-central-agent" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716683 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="ceilometer-central-agent" Oct 04 08:35:35 crc kubenswrapper[4969]: E1004 08:35:35.716694 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="ceilometer-notification-agent" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716701 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="ceilometer-notification-agent" Oct 04 08:35:35 crc kubenswrapper[4969]: E1004 08:35:35.716711 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff3f16f-8103-49a7-8a5c-07b6b7f399e7" containerName="mariadb-database-create" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716717 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff3f16f-8103-49a7-8a5c-07b6b7f399e7" containerName="mariadb-database-create" Oct 04 08:35:35 crc kubenswrapper[4969]: E1004 08:35:35.716734 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="proxy-httpd" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716740 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="proxy-httpd" Oct 04 08:35:35 crc kubenswrapper[4969]: E1004 08:35:35.716750 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="sg-core" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716756 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="sg-core" Oct 04 08:35:35 crc kubenswrapper[4969]: E1004 08:35:35.716769 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f" containerName="mariadb-database-create" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716775 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f" containerName="mariadb-database-create" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716946 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f" containerName="mariadb-database-create" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716962 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff3f16f-8103-49a7-8a5c-07b6b7f399e7" containerName="mariadb-database-create" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716969 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="ceilometer-central-agent" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716985 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="46a217c4-ea44-4608-9300-65410bec2b88" containerName="mariadb-database-create" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.716996 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="sg-core" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.717004 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="ceilometer-notification-agent" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.717018 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" containerName="proxy-httpd" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.718659 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.723240 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.723413 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.731313 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.863393 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.863751 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-log-httpd\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.863806 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwkw7\" (UniqueName: \"kubernetes.io/projected/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-kube-api-access-mwkw7\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.863895 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-run-httpd\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.863961 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-config-data\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.863993 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-scripts\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.864038 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.965999 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-config-data\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.966041 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-scripts\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.966073 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.966109 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.966203 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-log-httpd\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.966227 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwkw7\" (UniqueName: \"kubernetes.io/projected/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-kube-api-access-mwkw7\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.966252 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-run-httpd\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.966696 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-run-httpd\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.966799 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-log-httpd\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.970571 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.971197 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-config-data\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.973771 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-scripts\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.983799 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:35 crc kubenswrapper[4969]: I1004 08:35:35.985409 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwkw7\" (UniqueName: \"kubernetes.io/projected/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-kube-api-access-mwkw7\") pod \"ceilometer-0\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " pod="openstack/ceilometer-0" Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.048335 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.211612 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.212117 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cebd742d-94a0-4014-8172-7bf50d535215" containerName="glance-log" containerID="cri-o://faaf50efeefd4b36dea9c5e760a740a8eb39d398df5fc54ee80173f48c59d07f" gracePeriod=30 Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.212517 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cebd742d-94a0-4014-8172-7bf50d535215" containerName="glance-httpd" containerID="cri-o://7ac0f9aad61619cccff2e23bf891c80760c98c28eec34aa9f03324022d0bbb28" gracePeriod=30 Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.440089 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-75f87ddd45-2md8q" Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.579362 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.582852 4969 generic.go:334] "Generic (PLEG): container finished" podID="cebd742d-94a0-4014-8172-7bf50d535215" containerID="faaf50efeefd4b36dea9c5e760a740a8eb39d398df5fc54ee80173f48c59d07f" exitCode=143 Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.583488 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cebd742d-94a0-4014-8172-7bf50d535215","Type":"ContainerDied","Data":"faaf50efeefd4b36dea9c5e760a740a8eb39d398df5fc54ee80173f48c59d07f"} Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.981040 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.986183 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-db-sync-config-data\") pod \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.986327 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-etc-machine-id\") pod \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.986473 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st8n9\" (UniqueName: \"kubernetes.io/projected/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-kube-api-access-st8n9\") pod \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.986650 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-config-data\") pod \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.986687 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-combined-ca-bundle\") pod \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.986761 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-scripts\") pod \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\" (UID: \"278ebf83-6c66-4d97-8b5c-a7a83ace00b7\") " Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.987887 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "278ebf83-6c66-4d97-8b5c-a7a83ace00b7" (UID: "278ebf83-6c66-4d97-8b5c-a7a83ace00b7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:35:36 crc kubenswrapper[4969]: I1004 08:35:36.992841 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-kube-api-access-st8n9" (OuterVolumeSpecName: "kube-api-access-st8n9") pod "278ebf83-6c66-4d97-8b5c-a7a83ace00b7" (UID: "278ebf83-6c66-4d97-8b5c-a7a83ace00b7"). InnerVolumeSpecName "kube-api-access-st8n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.007777 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "278ebf83-6c66-4d97-8b5c-a7a83ace00b7" (UID: "278ebf83-6c66-4d97-8b5c-a7a83ace00b7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.049654 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-scripts" (OuterVolumeSpecName: "scripts") pod "278ebf83-6c66-4d97-8b5c-a7a83ace00b7" (UID: "278ebf83-6c66-4d97-8b5c-a7a83ace00b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.060492 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "278ebf83-6c66-4d97-8b5c-a7a83ace00b7" (UID: "278ebf83-6c66-4d97-8b5c-a7a83ace00b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.079804 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-config-data" (OuterVolumeSpecName: "config-data") pod "278ebf83-6c66-4d97-8b5c-a7a83ace00b7" (UID: "278ebf83-6c66-4d97-8b5c-a7a83ace00b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.091309 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.091341 4969 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.091354 4969 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.091363 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st8n9\" (UniqueName: \"kubernetes.io/projected/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-kube-api-access-st8n9\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.091375 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.091386 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/278ebf83-6c66-4d97-8b5c-a7a83ace00b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.124128 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="309264c4-c0cc-4ec7-9f90-6aed3a3379ce" path="/var/lib/kubelet/pods/309264c4-c0cc-4ec7-9f90-6aed3a3379ce/volumes" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.291620 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-74db-account-create-h78hj"] Oct 04 08:35:37 crc kubenswrapper[4969]: E1004 08:35:37.292844 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="278ebf83-6c66-4d97-8b5c-a7a83ace00b7" containerName="cinder-db-sync" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.292967 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="278ebf83-6c66-4d97-8b5c-a7a83ace00b7" containerName="cinder-db-sync" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.293142 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="278ebf83-6c66-4d97-8b5c-a7a83ace00b7" containerName="cinder-db-sync" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.293822 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-74db-account-create-h78hj" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.300449 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.317505 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-74db-account-create-h78hj"] Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.370556 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.396620 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8rss\" (UniqueName: \"kubernetes.io/projected/d4af2ba9-d233-4928-9a54-b5bf58ee50a9-kube-api-access-x8rss\") pod \"nova-api-74db-account-create-h78hj\" (UID: \"d4af2ba9-d233-4928-9a54-b5bf58ee50a9\") " pod="openstack/nova-api-74db-account-create-h78hj" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.429357 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d5ff7c687-t2ln2"] Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.429621 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" podUID="aaf3be84-c9a9-481c-8e10-a5dc928f9be9" containerName="dnsmasq-dns" containerID="cri-o://8163280de698d77ca2a24a53653e964d04932b8276ae73466ee2a11a1103a85c" gracePeriod=10 Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.496229 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-b89e-account-create-zlbmg"] Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.497594 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b89e-account-create-zlbmg" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.497633 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8rss\" (UniqueName: \"kubernetes.io/projected/d4af2ba9-d233-4928-9a54-b5bf58ee50a9-kube-api-access-x8rss\") pod \"nova-api-74db-account-create-h78hj\" (UID: \"d4af2ba9-d233-4928-9a54-b5bf58ee50a9\") " pod="openstack/nova-api-74db-account-create-h78hj" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.500114 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.523083 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8rss\" (UniqueName: \"kubernetes.io/projected/d4af2ba9-d233-4928-9a54-b5bf58ee50a9-kube-api-access-x8rss\") pod \"nova-api-74db-account-create-h78hj\" (UID: \"d4af2ba9-d233-4928-9a54-b5bf58ee50a9\") " pod="openstack/nova-api-74db-account-create-h78hj" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.524771 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b89e-account-create-zlbmg"] Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.598850 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99dsj\" (UniqueName: \"kubernetes.io/projected/00a41b61-3a97-4d3c-94b6-f76f5d58abe4-kube-api-access-99dsj\") pod \"nova-cell0-b89e-account-create-zlbmg\" (UID: \"00a41b61-3a97-4d3c-94b6-f76f5d58abe4\") " pod="openstack/nova-cell0-b89e-account-create-zlbmg" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.606873 4969 generic.go:334] "Generic (PLEG): container finished" podID="cebd742d-94a0-4014-8172-7bf50d535215" containerID="7ac0f9aad61619cccff2e23bf891c80760c98c28eec34aa9f03324022d0bbb28" exitCode=0 Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.606945 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cebd742d-94a0-4014-8172-7bf50d535215","Type":"ContainerDied","Data":"7ac0f9aad61619cccff2e23bf891c80760c98c28eec34aa9f03324022d0bbb28"} Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.608679 4969 generic.go:334] "Generic (PLEG): container finished" podID="aaf3be84-c9a9-481c-8e10-a5dc928f9be9" containerID="8163280de698d77ca2a24a53653e964d04932b8276ae73466ee2a11a1103a85c" exitCode=0 Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.608747 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" event={"ID":"aaf3be84-c9a9-481c-8e10-a5dc928f9be9","Type":"ContainerDied","Data":"8163280de698d77ca2a24a53653e964d04932b8276ae73466ee2a11a1103a85c"} Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.610452 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-74db-account-create-h78hj" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.610841 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerStarted","Data":"d780c380c1d32cebe53af088459c6341cadb1672c816fa07f3767231b8c3cea5"} Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.610868 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerStarted","Data":"0478a755f6cec6050cbf3952a3915eb4812cf2612e5e662909d2ec965716e6d5"} Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.610879 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerStarted","Data":"5f0d683aca406ef76ef49f5c985f33aa9b04a53a153bcfea6e364134ea0413cb"} Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.616642 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xjc7p" event={"ID":"278ebf83-6c66-4d97-8b5c-a7a83ace00b7","Type":"ContainerDied","Data":"16bf5704bca58400ccfe6aedaf4914de3e1bf935a3e155ade19202e72333893a"} Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.616678 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16bf5704bca58400ccfe6aedaf4914de3e1bf935a3e155ade19202e72333893a" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.616815 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xjc7p" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.703573 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99dsj\" (UniqueName: \"kubernetes.io/projected/00a41b61-3a97-4d3c-94b6-f76f5d58abe4-kube-api-access-99dsj\") pod \"nova-cell0-b89e-account-create-zlbmg\" (UID: \"00a41b61-3a97-4d3c-94b6-f76f5d58abe4\") " pod="openstack/nova-cell0-b89e-account-create-zlbmg" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.728000 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-56df-account-create-8jbkc"] Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.729273 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-56df-account-create-8jbkc" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.733477 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.734145 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99dsj\" (UniqueName: \"kubernetes.io/projected/00a41b61-3a97-4d3c-94b6-f76f5d58abe4-kube-api-access-99dsj\") pod \"nova-cell0-b89e-account-create-zlbmg\" (UID: \"00a41b61-3a97-4d3c-94b6-f76f5d58abe4\") " pod="openstack/nova-cell0-b89e-account-create-zlbmg" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.740467 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b89e-account-create-zlbmg" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.775825 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-56df-account-create-8jbkc"] Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.790861 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.837689 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 08:35:37 crc kubenswrapper[4969]: E1004 08:35:37.838436 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cebd742d-94a0-4014-8172-7bf50d535215" containerName="glance-log" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.838452 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="cebd742d-94a0-4014-8172-7bf50d535215" containerName="glance-log" Oct 04 08:35:37 crc kubenswrapper[4969]: E1004 08:35:37.838504 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cebd742d-94a0-4014-8172-7bf50d535215" containerName="glance-httpd" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.838512 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="cebd742d-94a0-4014-8172-7bf50d535215" containerName="glance-httpd" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.838718 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="cebd742d-94a0-4014-8172-7bf50d535215" containerName="glance-log" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.838747 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="cebd742d-94a0-4014-8172-7bf50d535215" containerName="glance-httpd" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.839925 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.843456 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xw5dg" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.843850 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.843966 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.844097 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.884783 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.906391 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-logs\") pod \"cebd742d-94a0-4014-8172-7bf50d535215\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.906573 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-public-tls-certs\") pod \"cebd742d-94a0-4014-8172-7bf50d535215\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.906607 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-scripts\") pod \"cebd742d-94a0-4014-8172-7bf50d535215\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.906654 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cebd742d-94a0-4014-8172-7bf50d535215\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.906726 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttz2n\" (UniqueName: \"kubernetes.io/projected/cebd742d-94a0-4014-8172-7bf50d535215-kube-api-access-ttz2n\") pod \"cebd742d-94a0-4014-8172-7bf50d535215\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.906806 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-httpd-run\") pod \"cebd742d-94a0-4014-8172-7bf50d535215\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.906864 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-config-data\") pod \"cebd742d-94a0-4014-8172-7bf50d535215\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.906899 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-combined-ca-bundle\") pod \"cebd742d-94a0-4014-8172-7bf50d535215\" (UID: \"cebd742d-94a0-4014-8172-7bf50d535215\") " Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.907199 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdgqz\" (UniqueName: \"kubernetes.io/projected/16495ba9-75b1-471b-8152-e477f327d94f-kube-api-access-kdgqz\") pod \"nova-cell1-56df-account-create-8jbkc\" (UID: \"16495ba9-75b1-471b-8152-e477f327d94f\") " pod="openstack/nova-cell1-56df-account-create-8jbkc" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.908177 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-logs" (OuterVolumeSpecName: "logs") pod "cebd742d-94a0-4014-8172-7bf50d535215" (UID: "cebd742d-94a0-4014-8172-7bf50d535215"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.908787 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cebd742d-94a0-4014-8172-7bf50d535215" (UID: "cebd742d-94a0-4014-8172-7bf50d535215"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.918978 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "cebd742d-94a0-4014-8172-7bf50d535215" (UID: "cebd742d-94a0-4014-8172-7bf50d535215"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.927799 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cebd742d-94a0-4014-8172-7bf50d535215-kube-api-access-ttz2n" (OuterVolumeSpecName: "kube-api-access-ttz2n") pod "cebd742d-94a0-4014-8172-7bf50d535215" (UID: "cebd742d-94a0-4014-8172-7bf50d535215"). InnerVolumeSpecName "kube-api-access-ttz2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.931231 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-scripts" (OuterVolumeSpecName: "scripts") pod "cebd742d-94a0-4014-8172-7bf50d535215" (UID: "cebd742d-94a0-4014-8172-7bf50d535215"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.940467 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b95fb5cb7-x5pbk"] Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.942040 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.955233 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b95fb5cb7-x5pbk"] Oct 04 08:35:37 crc kubenswrapper[4969]: I1004 08:35:37.992087 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cebd742d-94a0-4014-8172-7bf50d535215" (UID: "cebd742d-94a0-4014-8172-7bf50d535215"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.008957 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009025 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009046 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdgqz\" (UniqueName: \"kubernetes.io/projected/16495ba9-75b1-471b-8152-e477f327d94f-kube-api-access-kdgqz\") pod \"nova-cell1-56df-account-create-8jbkc\" (UID: \"16495ba9-75b1-471b-8152-e477f327d94f\") " pod="openstack/nova-cell1-56df-account-create-8jbkc" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009076 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009113 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-scripts\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009132 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79vjk\" (UniqueName: \"kubernetes.io/projected/612cc256-df44-4bf4-9c25-5a672fc14647-kube-api-access-79vjk\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009204 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/612cc256-df44-4bf4-9c25-5a672fc14647-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009251 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009260 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009269 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009286 4969 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009296 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttz2n\" (UniqueName: \"kubernetes.io/projected/cebd742d-94a0-4014-8172-7bf50d535215-kube-api-access-ttz2n\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.009305 4969 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cebd742d-94a0-4014-8172-7bf50d535215-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.037541 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cebd742d-94a0-4014-8172-7bf50d535215" (UID: "cebd742d-94a0-4014-8172-7bf50d535215"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.039651 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdgqz\" (UniqueName: \"kubernetes.io/projected/16495ba9-75b1-471b-8152-e477f327d94f-kube-api-access-kdgqz\") pod \"nova-cell1-56df-account-create-8jbkc\" (UID: \"16495ba9-75b1-471b-8152-e477f327d94f\") " pod="openstack/nova-cell1-56df-account-create-8jbkc" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.065552 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-config-data" (OuterVolumeSpecName: "config-data") pod "cebd742d-94a0-4014-8172-7bf50d535215" (UID: "cebd742d-94a0-4014-8172-7bf50d535215"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.067520 4969 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.084619 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-56df-account-create-8jbkc" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.100375 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.101872 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.105452 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111495 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/612cc256-df44-4bf4-9c25-5a672fc14647-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111530 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-sb\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111557 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-config\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111571 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/612cc256-df44-4bf4-9c25-5a672fc14647-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111593 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-nb\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111615 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111645 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-svc\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111682 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111710 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111730 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvlq9\" (UniqueName: \"kubernetes.io/projected/9b6a2e16-8be3-4946-b625-545f052e86f0-kube-api-access-wvlq9\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111766 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-scripts\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111787 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79vjk\" (UniqueName: \"kubernetes.io/projected/612cc256-df44-4bf4-9c25-5a672fc14647-kube-api-access-79vjk\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111825 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-swift-storage-0\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111892 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111905 4969 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebd742d-94a0-4014-8172-7bf50d535215-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.111915 4969 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.127843 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.128598 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.145467 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-scripts\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.146290 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.150022 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79vjk\" (UniqueName: \"kubernetes.io/projected/612cc256-df44-4bf4-9c25-5a672fc14647-kube-api-access-79vjk\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.160331 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data\") pod \"cinder-scheduler-0\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.178200 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.221794 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.221872 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-sb\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.221901 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-config\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.221937 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-nb\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.221960 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-svc\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.221994 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc18078e-24a4-4582-9d47-a5c0f4dced45-logs\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.222039 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvlq9\" (UniqueName: \"kubernetes.io/projected/9b6a2e16-8be3-4946-b625-545f052e86f0-kube-api-access-wvlq9\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.222078 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.222103 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-scripts\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.222125 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfbsf\" (UniqueName: \"kubernetes.io/projected/cc18078e-24a4-4582-9d47-a5c0f4dced45-kube-api-access-hfbsf\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.222142 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc18078e-24a4-4582-9d47-a5c0f4dced45-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.222183 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-swift-storage-0\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.222202 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data-custom\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.223393 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-sb\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.224133 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-config\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.225043 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-svc\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.225197 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-swift-storage-0\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.226708 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-nb\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.259556 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvlq9\" (UniqueName: \"kubernetes.io/projected/9b6a2e16-8be3-4946-b625-545f052e86f0-kube-api-access-wvlq9\") pod \"dnsmasq-dns-b95fb5cb7-x5pbk\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.265964 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.269742 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.333600 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc18078e-24a4-4582-9d47-a5c0f4dced45-logs\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.333969 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.334005 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-scripts\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.334037 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfbsf\" (UniqueName: \"kubernetes.io/projected/cc18078e-24a4-4582-9d47-a5c0f4dced45-kube-api-access-hfbsf\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.334057 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc18078e-24a4-4582-9d47-a5c0f4dced45-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.334083 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data-custom\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.334107 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.339040 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc18078e-24a4-4582-9d47-a5c0f4dced45-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.339837 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc18078e-24a4-4582-9d47-a5c0f4dced45-logs\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.346208 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-scripts\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.349200 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.360701 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data-custom\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.365781 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfbsf\" (UniqueName: \"kubernetes.io/projected/cc18078e-24a4-4582-9d47-a5c0f4dced45-kube-api-access-hfbsf\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.390712 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data\") pod \"cinder-api-0\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.435329 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-svc\") pod \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.435421 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-swift-storage-0\") pod \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.435477 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vk5w\" (UniqueName: \"kubernetes.io/projected/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-kube-api-access-5vk5w\") pod \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.435647 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-nb\") pod \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.435664 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-config\") pod \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.435705 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-sb\") pod \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\" (UID: \"aaf3be84-c9a9-481c-8e10-a5dc928f9be9\") " Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.453124 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-kube-api-access-5vk5w" (OuterVolumeSpecName: "kube-api-access-5vk5w") pod "aaf3be84-c9a9-481c-8e10-a5dc928f9be9" (UID: "aaf3be84-c9a9-481c-8e10-a5dc928f9be9"). InnerVolumeSpecName "kube-api-access-5vk5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.468977 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.493454 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-74db-account-create-h78hj"] Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.540358 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vk5w\" (UniqueName: \"kubernetes.io/projected/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-kube-api-access-5vk5w\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.586930 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b89e-account-create-zlbmg"] Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.619790 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aaf3be84-c9a9-481c-8e10-a5dc928f9be9" (UID: "aaf3be84-c9a9-481c-8e10-a5dc928f9be9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.628621 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aaf3be84-c9a9-481c-8e10-a5dc928f9be9" (UID: "aaf3be84-c9a9-481c-8e10-a5dc928f9be9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.656280 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.656313 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.665455 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-config" (OuterVolumeSpecName: "config") pod "aaf3be84-c9a9-481c-8e10-a5dc928f9be9" (UID: "aaf3be84-c9a9-481c-8e10-a5dc928f9be9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.677771 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aaf3be84-c9a9-481c-8e10-a5dc928f9be9" (UID: "aaf3be84-c9a9-481c-8e10-a5dc928f9be9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.713641 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aaf3be84-c9a9-481c-8e10-a5dc928f9be9" (UID: "aaf3be84-c9a9-481c-8e10-a5dc928f9be9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.746117 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" event={"ID":"aaf3be84-c9a9-481c-8e10-a5dc928f9be9","Type":"ContainerDied","Data":"680b10407f930ea70abcfd5bdcd4452f5afd5516364e2b64037c50e6c4407176"} Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.746173 4969 scope.go:117] "RemoveContainer" containerID="8163280de698d77ca2a24a53653e964d04932b8276ae73466ee2a11a1103a85c" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.746302 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5ff7c687-t2ln2" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.758172 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.759012 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.759024 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf3be84-c9a9-481c-8e10-a5dc928f9be9-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.782628 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerStarted","Data":"bacdaad22c40c28cb4fb83c9d1abc8a59669e2d6feb59081f9d9ec37ae87674b"} Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.788386 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-74db-account-create-h78hj" event={"ID":"d4af2ba9-d233-4928-9a54-b5bf58ee50a9","Type":"ContainerStarted","Data":"d9fa762e315a83bcbc935d6430a6682ff1beac6d9b728a6f8330d20eae06424a"} Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.809421 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d5ff7c687-t2ln2"] Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.809511 4969 scope.go:117] "RemoveContainer" containerID="989d9e3c86ede525cc5ffde902df95e5ef1fff7290f1cf9aac0416ee0bede741" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.819899 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cebd742d-94a0-4014-8172-7bf50d535215","Type":"ContainerDied","Data":"8111f2dd2b1bd14f9c2f06beca1ae39c0dd79be9cd7294e43a9324b2329f0875"} Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.820004 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.822419 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d5ff7c687-t2ln2"] Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.848893 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.921007 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.930436 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.944903 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:35:38 crc kubenswrapper[4969]: E1004 08:35:38.945328 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf3be84-c9a9-481c-8e10-a5dc928f9be9" containerName="dnsmasq-dns" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.945342 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf3be84-c9a9-481c-8e10-a5dc928f9be9" containerName="dnsmasq-dns" Oct 04 08:35:38 crc kubenswrapper[4969]: E1004 08:35:38.945352 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf3be84-c9a9-481c-8e10-a5dc928f9be9" containerName="init" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.945358 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf3be84-c9a9-481c-8e10-a5dc928f9be9" containerName="init" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.945552 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf3be84-c9a9-481c-8e10-a5dc928f9be9" containerName="dnsmasq-dns" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.946538 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.958248 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.958922 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.977620 4969 scope.go:117] "RemoveContainer" containerID="7ac0f9aad61619cccff2e23bf891c80760c98c28eec34aa9f03324022d0bbb28" Oct 04 08:35:38 crc kubenswrapper[4969]: I1004 08:35:38.981596 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.042482 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.044708 4969 scope.go:117] "RemoveContainer" containerID="faaf50efeefd4b36dea9c5e760a740a8eb39d398df5fc54ee80173f48c59d07f" Oct 04 08:35:39 crc kubenswrapper[4969]: W1004 08:35:39.045233 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod612cc256_df44_4bf4_9c25_5a672fc14647.slice/crio-9b055b417ebaff03faea4f0bd206f01823a7e1031b7eb5568e354b895beda565 WatchSource:0}: Error finding container 9b055b417ebaff03faea4f0bd206f01823a7e1031b7eb5568e354b895beda565: Status 404 returned error can't find the container with id 9b055b417ebaff03faea4f0bd206f01823a7e1031b7eb5568e354b895beda565 Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.068813 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.068865 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c627df9-6989-44d8-b72f-128992d65bd9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.068897 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-config-data\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.068940 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c627df9-6989-44d8-b72f-128992d65bd9-logs\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.068957 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.068994 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.069022 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n26b\" (UniqueName: \"kubernetes.io/projected/7c627df9-6989-44d8-b72f-128992d65bd9-kube-api-access-4n26b\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.069043 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-scripts\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.077159 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaf3be84-c9a9-481c-8e10-a5dc928f9be9" path="/var/lib/kubelet/pods/aaf3be84-c9a9-481c-8e10-a5dc928f9be9/volumes" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.077970 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cebd742d-94a0-4014-8172-7bf50d535215" path="/var/lib/kubelet/pods/cebd742d-94a0-4014-8172-7bf50d535215/volumes" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.169552 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-56df-account-create-8jbkc"] Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.170938 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.170996 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n26b\" (UniqueName: \"kubernetes.io/projected/7c627df9-6989-44d8-b72f-128992d65bd9-kube-api-access-4n26b\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.171020 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-scripts\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.171100 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.171143 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c627df9-6989-44d8-b72f-128992d65bd9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.171173 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-config-data\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.171234 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c627df9-6989-44d8-b72f-128992d65bd9-logs\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.171253 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.171755 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.171884 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c627df9-6989-44d8-b72f-128992d65bd9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.172480 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c627df9-6989-44d8-b72f-128992d65bd9-logs\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.178790 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.189522 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-scripts\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.197573 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n26b\" (UniqueName: \"kubernetes.io/projected/7c627df9-6989-44d8-b72f-128992d65bd9-kube-api-access-4n26b\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.202432 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-config-data\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.206283 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c627df9-6989-44d8-b72f-128992d65bd9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.240359 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"7c627df9-6989-44d8-b72f-128992d65bd9\") " pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.292740 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.322668 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.330544 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b95fb5cb7-x5pbk"] Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.434075 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.434509 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerName="glance-log" containerID="cri-o://aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0" gracePeriod=30 Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.434643 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerName="glance-httpd" containerID="cri-o://346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6" gracePeriod=30 Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.872912 4969 generic.go:334] "Generic (PLEG): container finished" podID="d4af2ba9-d233-4928-9a54-b5bf58ee50a9" containerID="56d1c29e6a36964868664598f9c3ae20edc0b600866ce852fdb6366ee59594f8" exitCode=0 Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.873225 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-74db-account-create-h78hj" event={"ID":"d4af2ba9-d233-4928-9a54-b5bf58ee50a9","Type":"ContainerDied","Data":"56d1c29e6a36964868664598f9c3ae20edc0b600866ce852fdb6366ee59594f8"} Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.895299 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-56df-account-create-8jbkc" event={"ID":"16495ba9-75b1-471b-8152-e477f327d94f","Type":"ContainerStarted","Data":"3ebf3b6ded4e06a3a91ac2f8c66d499f5b1001b8f789aea9e3211d7a7d85c31d"} Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.898027 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"612cc256-df44-4bf4-9c25-5a672fc14647","Type":"ContainerStarted","Data":"9b055b417ebaff03faea4f0bd206f01823a7e1031b7eb5568e354b895beda565"} Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.899271 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cc18078e-24a4-4582-9d47-a5c0f4dced45","Type":"ContainerStarted","Data":"afea206ea899ed91ad9e9648f67ab4fa53b2a5bd8fd56b2b516050f04a36eb72"} Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.901252 4969 generic.go:334] "Generic (PLEG): container finished" podID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerID="aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0" exitCode=143 Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.901354 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62","Type":"ContainerDied","Data":"aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0"} Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.904580 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" event={"ID":"9b6a2e16-8be3-4946-b625-545f052e86f0","Type":"ContainerStarted","Data":"683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502"} Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.904609 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" event={"ID":"9b6a2e16-8be3-4946-b625-545f052e86f0","Type":"ContainerStarted","Data":"98a8eebb705d7963f3fc6bb2ee15a275da93633596a6ceb8ef5da4693f9bd9b5"} Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.913139 4969 generic.go:334] "Generic (PLEG): container finished" podID="00a41b61-3a97-4d3c-94b6-f76f5d58abe4" containerID="89fbd59ae652ba88a603944657850a84aafc6ce9155b6c8fee9714ebe53fa82e" exitCode=0 Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.913191 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b89e-account-create-zlbmg" event={"ID":"00a41b61-3a97-4d3c-94b6-f76f5d58abe4","Type":"ContainerDied","Data":"89fbd59ae652ba88a603944657850a84aafc6ce9155b6c8fee9714ebe53fa82e"} Oct 04 08:35:39 crc kubenswrapper[4969]: I1004 08:35:39.913216 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b89e-account-create-zlbmg" event={"ID":"00a41b61-3a97-4d3c-94b6-f76f5d58abe4","Type":"ContainerStarted","Data":"4a75e35dd21fed39a8df8b1231782400a2ec5211502fde14cb8529fa0537a230"} Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.179855 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.440869 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.953889 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerStarted","Data":"c6ec224e27dcaed65bca9b38fb6ace57d9c17d80583020e6e2e5774381ab0905"} Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.954045 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="ceilometer-central-agent" containerID="cri-o://0478a755f6cec6050cbf3952a3915eb4812cf2612e5e662909d2ec965716e6d5" gracePeriod=30 Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.954239 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="proxy-httpd" containerID="cri-o://c6ec224e27dcaed65bca9b38fb6ace57d9c17d80583020e6e2e5774381ab0905" gracePeriod=30 Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.954276 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="ceilometer-notification-agent" containerID="cri-o://d780c380c1d32cebe53af088459c6341cadb1672c816fa07f3767231b8c3cea5" gracePeriod=30 Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.954322 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="sg-core" containerID="cri-o://bacdaad22c40c28cb4fb83c9d1abc8a59669e2d6feb59081f9d9ec37ae87674b" gracePeriod=30 Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.954372 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.965010 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7c627df9-6989-44d8-b72f-128992d65bd9","Type":"ContainerStarted","Data":"c743e53ea415249158d8be865c92297713021738851d72a38077f5058d7bf422"} Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.967636 4969 generic.go:334] "Generic (PLEG): container finished" podID="9b6a2e16-8be3-4946-b625-545f052e86f0" containerID="683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502" exitCode=0 Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.967697 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" event={"ID":"9b6a2e16-8be3-4946-b625-545f052e86f0","Type":"ContainerDied","Data":"683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502"} Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.967719 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" event={"ID":"9b6a2e16-8be3-4946-b625-545f052e86f0","Type":"ContainerStarted","Data":"75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa"} Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.968711 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.978766 4969 generic.go:334] "Generic (PLEG): container finished" podID="16495ba9-75b1-471b-8152-e477f327d94f" containerID="02584b220d350e940b2c33e9b081a7f4c3cceec5d5da78988633c34cfd0a3eab" exitCode=0 Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.978862 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-56df-account-create-8jbkc" event={"ID":"16495ba9-75b1-471b-8152-e477f327d94f","Type":"ContainerDied","Data":"02584b220d350e940b2c33e9b081a7f4c3cceec5d5da78988633c34cfd0a3eab"} Oct 04 08:35:40 crc kubenswrapper[4969]: I1004 08:35:40.984720 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.891557888 podStartE2EDuration="5.98470278s" podCreationTimestamp="2025-10-04 08:35:35 +0000 UTC" firstStartedPulling="2025-10-04 08:35:36.584835477 +0000 UTC m=+1164.339104291" lastFinishedPulling="2025-10-04 08:35:39.677980369 +0000 UTC m=+1167.432249183" observedRunningTime="2025-10-04 08:35:40.977561555 +0000 UTC m=+1168.731830389" watchObservedRunningTime="2025-10-04 08:35:40.98470278 +0000 UTC m=+1168.738971594" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.054748 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" podStartSLOduration=4.054731319 podStartE2EDuration="4.054731319s" podCreationTimestamp="2025-10-04 08:35:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:41.005128008 +0000 UTC m=+1168.759396822" watchObservedRunningTime="2025-10-04 08:35:41.054731319 +0000 UTC m=+1168.809000133" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.161162 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cc18078e-24a4-4582-9d47-a5c0f4dced45","Type":"ContainerStarted","Data":"17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0"} Oct 04 08:35:41 crc kubenswrapper[4969]: E1004 08:35:41.248920 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8d992f3_ca0e_4617_8e5d_500cc5184d2c.slice/crio-conmon-bacdaad22c40c28cb4fb83c9d1abc8a59669e2d6feb59081f9d9ec37ae87674b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4ae9c5e_7bfd_4fe7_b66f_2670589b7e62.slice/crio-346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4ae9c5e_7bfd_4fe7_b66f_2670589b7e62.slice/crio-conmon-346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8d992f3_ca0e_4617_8e5d_500cc5184d2c.slice/crio-bacdaad22c40c28cb4fb83c9d1abc8a59669e2d6feb59081f9d9ec37ae87674b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8d992f3_ca0e_4617_8e5d_500cc5184d2c.slice/crio-c6ec224e27dcaed65bca9b38fb6ace57d9c17d80583020e6e2e5774381ab0905.scope\": RecentStats: unable to find data in memory cache]" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.785201 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b89e-account-create-zlbmg" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.825633 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-74db-account-create-h78hj" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.865744 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8rss\" (UniqueName: \"kubernetes.io/projected/d4af2ba9-d233-4928-9a54-b5bf58ee50a9-kube-api-access-x8rss\") pod \"d4af2ba9-d233-4928-9a54-b5bf58ee50a9\" (UID: \"d4af2ba9-d233-4928-9a54-b5bf58ee50a9\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.865946 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99dsj\" (UniqueName: \"kubernetes.io/projected/00a41b61-3a97-4d3c-94b6-f76f5d58abe4-kube-api-access-99dsj\") pod \"00a41b61-3a97-4d3c-94b6-f76f5d58abe4\" (UID: \"00a41b61-3a97-4d3c-94b6-f76f5d58abe4\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.871167 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4af2ba9-d233-4928-9a54-b5bf58ee50a9-kube-api-access-x8rss" (OuterVolumeSpecName: "kube-api-access-x8rss") pod "d4af2ba9-d233-4928-9a54-b5bf58ee50a9" (UID: "d4af2ba9-d233-4928-9a54-b5bf58ee50a9"). InnerVolumeSpecName "kube-api-access-x8rss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.877842 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a41b61-3a97-4d3c-94b6-f76f5d58abe4-kube-api-access-99dsj" (OuterVolumeSpecName: "kube-api-access-99dsj") pod "00a41b61-3a97-4d3c-94b6-f76f5d58abe4" (UID: "00a41b61-3a97-4d3c-94b6-f76f5d58abe4"). InnerVolumeSpecName "kube-api-access-99dsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.907067 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.971854 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-config-data\") pod \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.971940 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-combined-ca-bundle\") pod \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.972086 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.972315 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-logs\") pod \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.972342 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-scripts\") pod \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.972383 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-internal-tls-certs\") pod \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.972529 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-httpd-run\") pod \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.972554 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4grdx\" (UniqueName: \"kubernetes.io/projected/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-kube-api-access-4grdx\") pod \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\" (UID: \"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62\") " Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.972934 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-logs" (OuterVolumeSpecName: "logs") pod "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" (UID: "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.972997 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8rss\" (UniqueName: \"kubernetes.io/projected/d4af2ba9-d233-4928-9a54-b5bf58ee50a9-kube-api-access-x8rss\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.973010 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99dsj\" (UniqueName: \"kubernetes.io/projected/00a41b61-3a97-4d3c-94b6-f76f5d58abe4-kube-api-access-99dsj\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.973276 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" (UID: "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.989975 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-kube-api-access-4grdx" (OuterVolumeSpecName: "kube-api-access-4grdx") pod "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" (UID: "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62"). InnerVolumeSpecName "kube-api-access-4grdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.990717 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" (UID: "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 08:35:41 crc kubenswrapper[4969]: I1004 08:35:41.997653 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-scripts" (OuterVolumeSpecName: "scripts") pod "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" (UID: "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.014812 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" (UID: "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.064546 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" (UID: "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.074683 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.074714 4969 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.074723 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.074732 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.074741 4969 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.074749 4969 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.074756 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4grdx\" (UniqueName: \"kubernetes.io/projected/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-kube-api-access-4grdx\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.081391 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b89e-account-create-zlbmg" event={"ID":"00a41b61-3a97-4d3c-94b6-f76f5d58abe4","Type":"ContainerDied","Data":"4a75e35dd21fed39a8df8b1231782400a2ec5211502fde14cb8529fa0537a230"} Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.081448 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a75e35dd21fed39a8df8b1231782400a2ec5211502fde14cb8529fa0537a230" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.081509 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b89e-account-create-zlbmg" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.085835 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"612cc256-df44-4bf4-9c25-5a672fc14647","Type":"ContainerStarted","Data":"b4336a0dad20763361d2da0cac58511b89a2ec064aedcf2d2acd8c75b06c7b1d"} Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.087557 4969 generic.go:334] "Generic (PLEG): container finished" podID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerID="c6ec224e27dcaed65bca9b38fb6ace57d9c17d80583020e6e2e5774381ab0905" exitCode=0 Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.087632 4969 generic.go:334] "Generic (PLEG): container finished" podID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerID="bacdaad22c40c28cb4fb83c9d1abc8a59669e2d6feb59081f9d9ec37ae87674b" exitCode=2 Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.087695 4969 generic.go:334] "Generic (PLEG): container finished" podID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerID="d780c380c1d32cebe53af088459c6341cadb1672c816fa07f3767231b8c3cea5" exitCode=0 Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.087768 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerDied","Data":"c6ec224e27dcaed65bca9b38fb6ace57d9c17d80583020e6e2e5774381ab0905"} Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.087831 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerDied","Data":"bacdaad22c40c28cb4fb83c9d1abc8a59669e2d6feb59081f9d9ec37ae87674b"} Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.087886 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerDied","Data":"d780c380c1d32cebe53af088459c6341cadb1672c816fa07f3767231b8c3cea5"} Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.088917 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-74db-account-create-h78hj" event={"ID":"d4af2ba9-d233-4928-9a54-b5bf58ee50a9","Type":"ContainerDied","Data":"d9fa762e315a83bcbc935d6430a6682ff1beac6d9b728a6f8330d20eae06424a"} Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.089001 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9fa762e315a83bcbc935d6430a6682ff1beac6d9b728a6f8330d20eae06424a" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.089107 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-74db-account-create-h78hj" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.095821 4969 generic.go:334] "Generic (PLEG): container finished" podID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerID="346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6" exitCode=0 Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.095879 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62","Type":"ContainerDied","Data":"346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6"} Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.095951 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62","Type":"ContainerDied","Data":"78279db8344044668848e865e24b5d5b6c610ca8fdfbfa28b6d35a32aed94673"} Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.096016 4969 scope.go:117] "RemoveContainer" containerID="346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.096185 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.115730 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7c627df9-6989-44d8-b72f-128992d65bd9","Type":"ContainerStarted","Data":"72fa957bc41458669cce928cad673cb53d269002d7ce9804fea21ccb2734f1e9"} Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.144312 4969 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.153719 4969 scope.go:117] "RemoveContainer" containerID="aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.164569 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-config-data" (OuterVolumeSpecName: "config-data") pod "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" (UID: "f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.177669 4969 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.177700 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.193765 4969 scope.go:117] "RemoveContainer" containerID="346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6" Oct 04 08:35:42 crc kubenswrapper[4969]: E1004 08:35:42.194246 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6\": container with ID starting with 346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6 not found: ID does not exist" containerID="346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.194273 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6"} err="failed to get container status \"346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6\": rpc error: code = NotFound desc = could not find container \"346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6\": container with ID starting with 346a30bb2656599db0ac713cd846d5b3ce8206ef17a0a7c941c9fd1a580ce5b6 not found: ID does not exist" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.194295 4969 scope.go:117] "RemoveContainer" containerID="aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0" Oct 04 08:35:42 crc kubenswrapper[4969]: E1004 08:35:42.194678 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0\": container with ID starting with aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0 not found: ID does not exist" containerID="aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.194697 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0"} err="failed to get container status \"aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0\": rpc error: code = NotFound desc = could not find container \"aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0\": container with ID starting with aa7bebd303ab1968a17fde9cece4a6e07041aa07fa352811cc9f5aabc0a0d6f0 not found: ID does not exist" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.480383 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.501996 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.523494 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:35:42 crc kubenswrapper[4969]: E1004 08:35:42.524020 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerName="glance-log" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.524043 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerName="glance-log" Oct 04 08:35:42 crc kubenswrapper[4969]: E1004 08:35:42.524055 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4af2ba9-d233-4928-9a54-b5bf58ee50a9" containerName="mariadb-account-create" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.524063 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4af2ba9-d233-4928-9a54-b5bf58ee50a9" containerName="mariadb-account-create" Oct 04 08:35:42 crc kubenswrapper[4969]: E1004 08:35:42.524073 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a41b61-3a97-4d3c-94b6-f76f5d58abe4" containerName="mariadb-account-create" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.524081 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a41b61-3a97-4d3c-94b6-f76f5d58abe4" containerName="mariadb-account-create" Oct 04 08:35:42 crc kubenswrapper[4969]: E1004 08:35:42.524112 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerName="glance-httpd" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.524121 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerName="glance-httpd" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.524376 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4af2ba9-d233-4928-9a54-b5bf58ee50a9" containerName="mariadb-account-create" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.524400 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a41b61-3a97-4d3c-94b6-f76f5d58abe4" containerName="mariadb-account-create" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.524423 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerName="glance-httpd" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.524513 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" containerName="glance-log" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.530200 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.544648 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.544838 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.544861 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.655053 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-56df-account-create-8jbkc" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.687609 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.687704 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-logs\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.687744 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8jlj\" (UniqueName: \"kubernetes.io/projected/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-kube-api-access-r8jlj\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.687943 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.688046 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.688132 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.688231 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.688275 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.789607 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdgqz\" (UniqueName: \"kubernetes.io/projected/16495ba9-75b1-471b-8152-e477f327d94f-kube-api-access-kdgqz\") pod \"16495ba9-75b1-471b-8152-e477f327d94f\" (UID: \"16495ba9-75b1-471b-8152-e477f327d94f\") " Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.790000 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.790057 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.790084 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.790143 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.790174 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.790197 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.790225 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-logs\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.790246 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8jlj\" (UniqueName: \"kubernetes.io/projected/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-kube-api-access-r8jlj\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.791567 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.791813 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-logs\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.792194 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.801623 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.802520 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.802617 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.803556 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.804492 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16495ba9-75b1-471b-8152-e477f327d94f-kube-api-access-kdgqz" (OuterVolumeSpecName: "kube-api-access-kdgqz") pod "16495ba9-75b1-471b-8152-e477f327d94f" (UID: "16495ba9-75b1-471b-8152-e477f327d94f"). InnerVolumeSpecName "kube-api-access-kdgqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.807131 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8jlj\" (UniqueName: \"kubernetes.io/projected/7ebb3f7a-584d-4c2d-a26f-f66481883b6e-kube-api-access-r8jlj\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.831812 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7ebb3f7a-584d-4c2d-a26f-f66481883b6e\") " pod="openstack/glance-default-internal-api-0" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.893710 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdgqz\" (UniqueName: \"kubernetes.io/projected/16495ba9-75b1-471b-8152-e477f327d94f-kube-api-access-kdgqz\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:42 crc kubenswrapper[4969]: I1004 08:35:42.952433 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.068352 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62" path="/var/lib/kubelet/pods/f4ae9c5e-7bfd-4fe7-b66f-2670589b7e62/volumes" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.131966 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"612cc256-df44-4bf4-9c25-5a672fc14647","Type":"ContainerStarted","Data":"e07ee3d3d16b99fa24c649e0da23a789abdff7ab8cfba7afbbcd530ce1ca385a"} Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.142968 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cc18078e-24a4-4582-9d47-a5c0f4dced45","Type":"ContainerStarted","Data":"031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f"} Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.143345 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerName="cinder-api-log" containerID="cri-o://17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0" gracePeriod=30 Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.143513 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.143559 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerName="cinder-api" containerID="cri-o://031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f" gracePeriod=30 Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.160105 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7c627df9-6989-44d8-b72f-128992d65bd9","Type":"ContainerStarted","Data":"9a36c2a0146bd401369cfa2c010bb5ef3899a945350fedc10bdef5245502bf01"} Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.168152 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-56df-account-create-8jbkc" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.168991 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-56df-account-create-8jbkc" event={"ID":"16495ba9-75b1-471b-8152-e477f327d94f","Type":"ContainerDied","Data":"3ebf3b6ded4e06a3a91ac2f8c66d499f5b1001b8f789aea9e3211d7a7d85c31d"} Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.169088 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ebf3b6ded4e06a3a91ac2f8c66d499f5b1001b8f789aea9e3211d7a7d85c31d" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.178627 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.203165 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.796190452 podStartE2EDuration="6.203147587s" podCreationTimestamp="2025-10-04 08:35:37 +0000 UTC" firstStartedPulling="2025-10-04 08:35:39.051565546 +0000 UTC m=+1166.805834360" lastFinishedPulling="2025-10-04 08:35:39.458522681 +0000 UTC m=+1167.212791495" observedRunningTime="2025-10-04 08:35:43.154293714 +0000 UTC m=+1170.908562528" watchObservedRunningTime="2025-10-04 08:35:43.203147587 +0000 UTC m=+1170.957416401" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.212145 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.212125747 podStartE2EDuration="5.212125747s" podCreationTimestamp="2025-10-04 08:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:43.178943187 +0000 UTC m=+1170.933212001" watchObservedRunningTime="2025-10-04 08:35:43.212125747 +0000 UTC m=+1170.966394561" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.221367 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.221349642 podStartE2EDuration="5.221349642s" podCreationTimestamp="2025-10-04 08:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:43.202660516 +0000 UTC m=+1170.956929330" watchObservedRunningTime="2025-10-04 08:35:43.221349642 +0000 UTC m=+1170.975618456" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.529160 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.786900 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.917968 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-scripts\") pod \"cc18078e-24a4-4582-9d47-a5c0f4dced45\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.918106 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-combined-ca-bundle\") pod \"cc18078e-24a4-4582-9d47-a5c0f4dced45\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.918199 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfbsf\" (UniqueName: \"kubernetes.io/projected/cc18078e-24a4-4582-9d47-a5c0f4dced45-kube-api-access-hfbsf\") pod \"cc18078e-24a4-4582-9d47-a5c0f4dced45\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.918765 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data\") pod \"cc18078e-24a4-4582-9d47-a5c0f4dced45\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.918844 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc18078e-24a4-4582-9d47-a5c0f4dced45-etc-machine-id\") pod \"cc18078e-24a4-4582-9d47-a5c0f4dced45\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.918895 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc18078e-24a4-4582-9d47-a5c0f4dced45-logs\") pod \"cc18078e-24a4-4582-9d47-a5c0f4dced45\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.918922 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data-custom\") pod \"cc18078e-24a4-4582-9d47-a5c0f4dced45\" (UID: \"cc18078e-24a4-4582-9d47-a5c0f4dced45\") " Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.918962 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc18078e-24a4-4582-9d47-a5c0f4dced45-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cc18078e-24a4-4582-9d47-a5c0f4dced45" (UID: "cc18078e-24a4-4582-9d47-a5c0f4dced45"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.919227 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc18078e-24a4-4582-9d47-a5c0f4dced45-logs" (OuterVolumeSpecName: "logs") pod "cc18078e-24a4-4582-9d47-a5c0f4dced45" (UID: "cc18078e-24a4-4582-9d47-a5c0f4dced45"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.919668 4969 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc18078e-24a4-4582-9d47-a5c0f4dced45-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.919686 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc18078e-24a4-4582-9d47-a5c0f4dced45-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.923438 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc18078e-24a4-4582-9d47-a5c0f4dced45-kube-api-access-hfbsf" (OuterVolumeSpecName: "kube-api-access-hfbsf") pod "cc18078e-24a4-4582-9d47-a5c0f4dced45" (UID: "cc18078e-24a4-4582-9d47-a5c0f4dced45"). InnerVolumeSpecName "kube-api-access-hfbsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.923656 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cc18078e-24a4-4582-9d47-a5c0f4dced45" (UID: "cc18078e-24a4-4582-9d47-a5c0f4dced45"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.923828 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-scripts" (OuterVolumeSpecName: "scripts") pod "cc18078e-24a4-4582-9d47-a5c0f4dced45" (UID: "cc18078e-24a4-4582-9d47-a5c0f4dced45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.955473 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc18078e-24a4-4582-9d47-a5c0f4dced45" (UID: "cc18078e-24a4-4582-9d47-a5c0f4dced45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:43 crc kubenswrapper[4969]: I1004 08:35:43.987670 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data" (OuterVolumeSpecName: "config-data") pod "cc18078e-24a4-4582-9d47-a5c0f4dced45" (UID: "cc18078e-24a4-4582-9d47-a5c0f4dced45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.021460 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.021488 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfbsf\" (UniqueName: \"kubernetes.io/projected/cc18078e-24a4-4582-9d47-a5c0f4dced45-kube-api-access-hfbsf\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.021499 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.021508 4969 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.021516 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc18078e-24a4-4582-9d47-a5c0f4dced45-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.194789 4969 generic.go:334] "Generic (PLEG): container finished" podID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerID="031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f" exitCode=0 Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.195306 4969 generic.go:334] "Generic (PLEG): container finished" podID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerID="17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0" exitCode=143 Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.195346 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cc18078e-24a4-4582-9d47-a5c0f4dced45","Type":"ContainerDied","Data":"031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f"} Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.195374 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cc18078e-24a4-4582-9d47-a5c0f4dced45","Type":"ContainerDied","Data":"17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0"} Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.195385 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cc18078e-24a4-4582-9d47-a5c0f4dced45","Type":"ContainerDied","Data":"afea206ea899ed91ad9e9648f67ab4fa53b2a5bd8fd56b2b516050f04a36eb72"} Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.195399 4969 scope.go:117] "RemoveContainer" containerID="031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.195572 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.209338 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7ebb3f7a-584d-4c2d-a26f-f66481883b6e","Type":"ContainerStarted","Data":"7652c807a056c9259c93f405c3cd26689ebdbf66f2dfbcf4ba4678576a181be8"} Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.209405 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7ebb3f7a-584d-4c2d-a26f-f66481883b6e","Type":"ContainerStarted","Data":"70bfb0790f423fdb051252c9892a0154bcc0e4d7dab4531c98ea48acda144f4f"} Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.231503 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.244500 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.251003 4969 scope.go:117] "RemoveContainer" containerID="17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.255562 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 08:35:44 crc kubenswrapper[4969]: E1004 08:35:44.255986 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16495ba9-75b1-471b-8152-e477f327d94f" containerName="mariadb-account-create" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.256010 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="16495ba9-75b1-471b-8152-e477f327d94f" containerName="mariadb-account-create" Oct 04 08:35:44 crc kubenswrapper[4969]: E1004 08:35:44.256031 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerName="cinder-api" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.256040 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerName="cinder-api" Oct 04 08:35:44 crc kubenswrapper[4969]: E1004 08:35:44.256055 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerName="cinder-api-log" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.256062 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerName="cinder-api-log" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.256244 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerName="cinder-api" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.256262 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc18078e-24a4-4582-9d47-a5c0f4dced45" containerName="cinder-api-log" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.256274 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="16495ba9-75b1-471b-8152-e477f327d94f" containerName="mariadb-account-create" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.257299 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.265038 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.265148 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.265182 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.275557 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.304920 4969 scope.go:117] "RemoveContainer" containerID="031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f" Oct 04 08:35:44 crc kubenswrapper[4969]: E1004 08:35:44.306123 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f\": container with ID starting with 031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f not found: ID does not exist" containerID="031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.306196 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f"} err="failed to get container status \"031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f\": rpc error: code = NotFound desc = could not find container \"031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f\": container with ID starting with 031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f not found: ID does not exist" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.306614 4969 scope.go:117] "RemoveContainer" containerID="17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0" Oct 04 08:35:44 crc kubenswrapper[4969]: E1004 08:35:44.306906 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0\": container with ID starting with 17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0 not found: ID does not exist" containerID="17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.306942 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0"} err="failed to get container status \"17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0\": rpc error: code = NotFound desc = could not find container \"17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0\": container with ID starting with 17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0 not found: ID does not exist" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.306967 4969 scope.go:117] "RemoveContainer" containerID="031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.308718 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f"} err="failed to get container status \"031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f\": rpc error: code = NotFound desc = could not find container \"031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f\": container with ID starting with 031f26813712c5d4782e463b2a2ec4b3c4457c35e6f46bfc59f5e6ad4b5e2b1f not found: ID does not exist" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.308739 4969 scope.go:117] "RemoveContainer" containerID="17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.309100 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0"} err="failed to get container status \"17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0\": rpc error: code = NotFound desc = could not find container \"17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0\": container with ID starting with 17729fb6113b4eeb9f7402292e0f17c00623e4e2a5a94f6d8f3a9ea9d071caf0 not found: ID does not exist" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.327879 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-config-data-custom\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.327959 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-config-data\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.328022 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.328050 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.328130 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.328223 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755c8160-b3df-4ba0-8d91-46d4af514dfc-logs\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.328278 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/755c8160-b3df-4ba0-8d91-46d4af514dfc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.328296 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-scripts\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.328405 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w49pd\" (UniqueName: \"kubernetes.io/projected/755c8160-b3df-4ba0-8d91-46d4af514dfc-kube-api-access-w49pd\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.429880 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w49pd\" (UniqueName: \"kubernetes.io/projected/755c8160-b3df-4ba0-8d91-46d4af514dfc-kube-api-access-w49pd\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.430022 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-config-data-custom\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.430083 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-config-data\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.430128 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.430174 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.430256 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.430315 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755c8160-b3df-4ba0-8d91-46d4af514dfc-logs\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.430366 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/755c8160-b3df-4ba0-8d91-46d4af514dfc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.430403 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-scripts\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.431907 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755c8160-b3df-4ba0-8d91-46d4af514dfc-logs\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.433631 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/755c8160-b3df-4ba0-8d91-46d4af514dfc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.434114 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.434830 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.436273 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-config-data-custom\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.442104 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.442294 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-config-data\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.442448 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/755c8160-b3df-4ba0-8d91-46d4af514dfc-scripts\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.451112 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w49pd\" (UniqueName: \"kubernetes.io/projected/755c8160-b3df-4ba0-8d91-46d4af514dfc-kube-api-access-w49pd\") pod \"cinder-api-0\" (UID: \"755c8160-b3df-4ba0-8d91-46d4af514dfc\") " pod="openstack/cinder-api-0" Oct 04 08:35:44 crc kubenswrapper[4969]: I1004 08:35:44.597728 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.065658 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc18078e-24a4-4582-9d47-a5c0f4dced45" path="/var/lib/kubelet/pods/cc18078e-24a4-4582-9d47-a5c0f4dced45/volumes" Oct 04 08:35:45 crc kubenswrapper[4969]: W1004 08:35:45.066151 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod755c8160_b3df_4ba0_8d91_46d4af514dfc.slice/crio-ae23810996ad28ce06e4c7894e34aa5c2fb7032998ea9e4edca42dee3b023edb WatchSource:0}: Error finding container ae23810996ad28ce06e4c7894e34aa5c2fb7032998ea9e4edca42dee3b023edb: Status 404 returned error can't find the container with id ae23810996ad28ce06e4c7894e34aa5c2fb7032998ea9e4edca42dee3b023edb Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.071122 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.229776 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7ebb3f7a-584d-4c2d-a26f-f66481883b6e","Type":"ContainerStarted","Data":"567c5793217ffe80cdc180c060db6b4d763051203e109545a5b79f2bc2816ae3"} Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.254204 4969 generic.go:334] "Generic (PLEG): container finished" podID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerID="0478a755f6cec6050cbf3952a3915eb4812cf2612e5e662909d2ec965716e6d5" exitCode=0 Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.254292 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerDied","Data":"0478a755f6cec6050cbf3952a3915eb4812cf2612e5e662909d2ec965716e6d5"} Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.257713 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.257689925 podStartE2EDuration="3.257689925s" podCreationTimestamp="2025-10-04 08:35:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:45.247070845 +0000 UTC m=+1173.001339669" watchObservedRunningTime="2025-10-04 08:35:45.257689925 +0000 UTC m=+1173.011958739" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.262014 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"755c8160-b3df-4ba0-8d91-46d4af514dfc","Type":"ContainerStarted","Data":"ae23810996ad28ce06e4c7894e34aa5c2fb7032998ea9e4edca42dee3b023edb"} Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.657030 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.755148 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-sg-core-conf-yaml\") pod \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.755317 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-run-httpd\") pod \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.755353 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwkw7\" (UniqueName: \"kubernetes.io/projected/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-kube-api-access-mwkw7\") pod \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.755483 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-combined-ca-bundle\") pod \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.755512 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-scripts\") pod \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.755532 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-config-data\") pod \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.755568 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-log-httpd\") pod \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\" (UID: \"c8d992f3-ca0e-4617-8e5d-500cc5184d2c\") " Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.756160 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c8d992f3-ca0e-4617-8e5d-500cc5184d2c" (UID: "c8d992f3-ca0e-4617-8e5d-500cc5184d2c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.756406 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c8d992f3-ca0e-4617-8e5d-500cc5184d2c" (UID: "c8d992f3-ca0e-4617-8e5d-500cc5184d2c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.773953 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-scripts" (OuterVolumeSpecName: "scripts") pod "c8d992f3-ca0e-4617-8e5d-500cc5184d2c" (UID: "c8d992f3-ca0e-4617-8e5d-500cc5184d2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.783532 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-kube-api-access-mwkw7" (OuterVolumeSpecName: "kube-api-access-mwkw7") pod "c8d992f3-ca0e-4617-8e5d-500cc5184d2c" (UID: "c8d992f3-ca0e-4617-8e5d-500cc5184d2c"). InnerVolumeSpecName "kube-api-access-mwkw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.847689 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c8d992f3-ca0e-4617-8e5d-500cc5184d2c" (UID: "c8d992f3-ca0e-4617-8e5d-500cc5184d2c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.855588 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8d992f3-ca0e-4617-8e5d-500cc5184d2c" (UID: "c8d992f3-ca0e-4617-8e5d-500cc5184d2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.858789 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.858832 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.858845 4969 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.858858 4969 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.858869 4969 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.858883 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwkw7\" (UniqueName: \"kubernetes.io/projected/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-kube-api-access-mwkw7\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.879843 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-config-data" (OuterVolumeSpecName: "config-data") pod "c8d992f3-ca0e-4617-8e5d-500cc5184d2c" (UID: "c8d992f3-ca0e-4617-8e5d-500cc5184d2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:45 crc kubenswrapper[4969]: I1004 08:35:45.959936 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d992f3-ca0e-4617-8e5d-500cc5184d2c-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.288693 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8d992f3-ca0e-4617-8e5d-500cc5184d2c","Type":"ContainerDied","Data":"5f0d683aca406ef76ef49f5c985f33aa9b04a53a153bcfea6e364134ea0413cb"} Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.288734 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.288743 4969 scope.go:117] "RemoveContainer" containerID="c6ec224e27dcaed65bca9b38fb6ace57d9c17d80583020e6e2e5774381ab0905" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.298103 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"755c8160-b3df-4ba0-8d91-46d4af514dfc","Type":"ContainerStarted","Data":"74d55169bdd88dd724812239a802815f8f23f6768029f1b4d7a434076f61472a"} Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.310502 4969 scope.go:117] "RemoveContainer" containerID="bacdaad22c40c28cb4fb83c9d1abc8a59669e2d6feb59081f9d9ec37ae87674b" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.343795 4969 scope.go:117] "RemoveContainer" containerID="d780c380c1d32cebe53af088459c6341cadb1672c816fa07f3767231b8c3cea5" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.349361 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.371779 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.384309 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:46 crc kubenswrapper[4969]: E1004 08:35:46.384918 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="sg-core" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.385348 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="sg-core" Oct 04 08:35:46 crc kubenswrapper[4969]: E1004 08:35:46.385982 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="ceilometer-central-agent" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.386094 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="ceilometer-central-agent" Oct 04 08:35:46 crc kubenswrapper[4969]: E1004 08:35:46.386170 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="proxy-httpd" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.386226 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="proxy-httpd" Oct 04 08:35:46 crc kubenswrapper[4969]: E1004 08:35:46.386292 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="ceilometer-notification-agent" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.386341 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="ceilometer-notification-agent" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.386711 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="ceilometer-central-agent" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.386791 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="proxy-httpd" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.386861 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="sg-core" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.386921 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" containerName="ceilometer-notification-agent" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.393590 4969 scope.go:117] "RemoveContainer" containerID="0478a755f6cec6050cbf3952a3915eb4812cf2612e5e662909d2ec965716e6d5" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.394585 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.398016 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.402067 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.404733 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.441615 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:46 crc kubenswrapper[4969]: E1004 08:35:46.455773 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-g8jgt log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data kube-api-access-g8jgt log-httpd run-httpd scripts sg-core-conf-yaml]: context canceled" pod="openstack/ceilometer-0" podUID="0167a1f5-d1bd-4878-823a-b45304df703a" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.481393 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.481471 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-config-data\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.481501 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.481560 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8jgt\" (UniqueName: \"kubernetes.io/projected/0167a1f5-d1bd-4878-823a-b45304df703a-kube-api-access-g8jgt\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.482145 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-run-httpd\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.482194 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-log-httpd\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.482276 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-scripts\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.584659 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-run-httpd\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.584700 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-log-httpd\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.584750 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-scripts\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.584831 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.584975 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-config-data\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.585395 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-log-httpd\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.585413 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.585568 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8jgt\" (UniqueName: \"kubernetes.io/projected/0167a1f5-d1bd-4878-823a-b45304df703a-kube-api-access-g8jgt\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.585834 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-run-httpd\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.589164 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.589700 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.590255 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-config-data\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.590691 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-scripts\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:46 crc kubenswrapper[4969]: I1004 08:35:46.604999 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8jgt\" (UniqueName: \"kubernetes.io/projected/0167a1f5-d1bd-4878-823a-b45304df703a-kube-api-access-g8jgt\") pod \"ceilometer-0\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " pod="openstack/ceilometer-0" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.056514 4969 scope.go:117] "RemoveContainer" containerID="84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58" Oct 04 08:35:47 crc kubenswrapper[4969]: E1004 08:35:47.057066 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.066942 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8d992f3-ca0e-4617-8e5d-500cc5184d2c" path="/var/lib/kubelet/pods/c8d992f3-ca0e-4617-8e5d-500cc5184d2c/volumes" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.309561 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"755c8160-b3df-4ba0-8d91-46d4af514dfc","Type":"ContainerStarted","Data":"6a12de7b17eeca278b8128fe642675c0c9c4998e6e878f360815b7cc651eb4d0"} Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.310573 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.311991 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.325291 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.403473 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-sg-core-conf-yaml\") pod \"0167a1f5-d1bd-4878-823a-b45304df703a\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.403666 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8jgt\" (UniqueName: \"kubernetes.io/projected/0167a1f5-d1bd-4878-823a-b45304df703a-kube-api-access-g8jgt\") pod \"0167a1f5-d1bd-4878-823a-b45304df703a\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.403738 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-combined-ca-bundle\") pod \"0167a1f5-d1bd-4878-823a-b45304df703a\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.403834 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-scripts\") pod \"0167a1f5-d1bd-4878-823a-b45304df703a\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.403876 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-log-httpd\") pod \"0167a1f5-d1bd-4878-823a-b45304df703a\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.403936 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-config-data\") pod \"0167a1f5-d1bd-4878-823a-b45304df703a\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.403982 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-run-httpd\") pod \"0167a1f5-d1bd-4878-823a-b45304df703a\" (UID: \"0167a1f5-d1bd-4878-823a-b45304df703a\") " Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.404141 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0167a1f5-d1bd-4878-823a-b45304df703a" (UID: "0167a1f5-d1bd-4878-823a-b45304df703a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.404411 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0167a1f5-d1bd-4878-823a-b45304df703a" (UID: "0167a1f5-d1bd-4878-823a-b45304df703a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.404939 4969 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.404993 4969 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0167a1f5-d1bd-4878-823a-b45304df703a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.407640 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0167a1f5-d1bd-4878-823a-b45304df703a" (UID: "0167a1f5-d1bd-4878-823a-b45304df703a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.409792 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-scripts" (OuterVolumeSpecName: "scripts") pod "0167a1f5-d1bd-4878-823a-b45304df703a" (UID: "0167a1f5-d1bd-4878-823a-b45304df703a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.410181 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-config-data" (OuterVolumeSpecName: "config-data") pod "0167a1f5-d1bd-4878-823a-b45304df703a" (UID: "0167a1f5-d1bd-4878-823a-b45304df703a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.411470 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0167a1f5-d1bd-4878-823a-b45304df703a-kube-api-access-g8jgt" (OuterVolumeSpecName: "kube-api-access-g8jgt") pod "0167a1f5-d1bd-4878-823a-b45304df703a" (UID: "0167a1f5-d1bd-4878-823a-b45304df703a"). InnerVolumeSpecName "kube-api-access-g8jgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.411537 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0167a1f5-d1bd-4878-823a-b45304df703a" (UID: "0167a1f5-d1bd-4878-823a-b45304df703a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.506115 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8jgt\" (UniqueName: \"kubernetes.io/projected/0167a1f5-d1bd-4878-823a-b45304df703a-kube-api-access-g8jgt\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.506156 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.506165 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.506174 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.506182 4969 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0167a1f5-d1bd-4878-823a-b45304df703a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.884848 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.88482914 podStartE2EDuration="3.88482914s" podCreationTimestamp="2025-10-04 08:35:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:35:47.335768466 +0000 UTC m=+1175.090037290" watchObservedRunningTime="2025-10-04 08:35:47.88482914 +0000 UTC m=+1175.639097954" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.889470 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjwv2"] Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.890632 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.895854 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.896063 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.896088 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bczdv" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.899292 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjwv2"] Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.924319 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-config-data\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.924406 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-scripts\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.924504 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp9s2\" (UniqueName: \"kubernetes.io/projected/8a0efc12-8421-489a-8fe9-97ae54119cc8-kube-api-access-dp9s2\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:47 crc kubenswrapper[4969]: I1004 08:35:47.924546 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.026145 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-scripts\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.026253 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp9s2\" (UniqueName: \"kubernetes.io/projected/8a0efc12-8421-489a-8fe9-97ae54119cc8-kube-api-access-dp9s2\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.026302 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.026349 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-config-data\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.031803 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-config-data\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.033415 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-scripts\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.033960 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.052164 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp9s2\" (UniqueName: \"kubernetes.io/projected/8a0efc12-8421-489a-8fe9-97ae54119cc8-kube-api-access-dp9s2\") pod \"nova-cell0-conductor-db-sync-pjwv2\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.206534 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.272587 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.326606 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.368751 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b8d76ff-7kf7d"] Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.370269 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" podUID="6a27a749-7256-45fc-b79c-4882d5554e9d" containerName="dnsmasq-dns" containerID="cri-o://ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b" gracePeriod=10 Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.413488 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.418530 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.432129 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.436210 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.458083 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.463037 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.474189 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.474374 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.525467 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.535145 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-config-data\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.535597 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-log-httpd\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.535627 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-run-httpd\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.535671 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-scripts\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.535737 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4r8h\" (UniqueName: \"kubernetes.io/projected/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-kube-api-access-d4r8h\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.535787 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.535834 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.636386 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-scripts\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.636484 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4r8h\" (UniqueName: \"kubernetes.io/projected/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-kube-api-access-d4r8h\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.636520 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.636535 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.636582 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-config-data\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.636603 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-log-httpd\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.636627 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-run-httpd\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.637546 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-run-httpd\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.637745 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-log-httpd\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.644926 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.647422 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-scripts\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.653860 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4r8h\" (UniqueName: \"kubernetes.io/projected/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-kube-api-access-d4r8h\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.656846 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-config-data\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.667049 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.731121 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjwv2"] Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.870086 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:35:48 crc kubenswrapper[4969]: I1004 08:35:48.981321 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.051436 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-nb\") pod \"6a27a749-7256-45fc-b79c-4882d5554e9d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.051738 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-swift-storage-0\") pod \"6a27a749-7256-45fc-b79c-4882d5554e9d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.051820 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-sb\") pod \"6a27a749-7256-45fc-b79c-4882d5554e9d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.051971 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8np22\" (UniqueName: \"kubernetes.io/projected/6a27a749-7256-45fc-b79c-4882d5554e9d-kube-api-access-8np22\") pod \"6a27a749-7256-45fc-b79c-4882d5554e9d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.052195 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-config\") pod \"6a27a749-7256-45fc-b79c-4882d5554e9d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.052437 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-svc\") pod \"6a27a749-7256-45fc-b79c-4882d5554e9d\" (UID: \"6a27a749-7256-45fc-b79c-4882d5554e9d\") " Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.074939 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a27a749-7256-45fc-b79c-4882d5554e9d-kube-api-access-8np22" (OuterVolumeSpecName: "kube-api-access-8np22") pod "6a27a749-7256-45fc-b79c-4882d5554e9d" (UID: "6a27a749-7256-45fc-b79c-4882d5554e9d"). InnerVolumeSpecName "kube-api-access-8np22". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.117780 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0167a1f5-d1bd-4878-823a-b45304df703a" path="/var/lib/kubelet/pods/0167a1f5-d1bd-4878-823a-b45304df703a/volumes" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.132666 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a27a749-7256-45fc-b79c-4882d5554e9d" (UID: "6a27a749-7256-45fc-b79c-4882d5554e9d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.154105 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6a27a749-7256-45fc-b79c-4882d5554e9d" (UID: "6a27a749-7256-45fc-b79c-4882d5554e9d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.159857 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.159885 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.160020 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8np22\" (UniqueName: \"kubernetes.io/projected/6a27a749-7256-45fc-b79c-4882d5554e9d-kube-api-access-8np22\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.166626 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-config" (OuterVolumeSpecName: "config") pod "6a27a749-7256-45fc-b79c-4882d5554e9d" (UID: "6a27a749-7256-45fc-b79c-4882d5554e9d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.168320 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6a27a749-7256-45fc-b79c-4882d5554e9d" (UID: "6a27a749-7256-45fc-b79c-4882d5554e9d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.170383 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6a27a749-7256-45fc-b79c-4882d5554e9d" (UID: "6a27a749-7256-45fc-b79c-4882d5554e9d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.262684 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.262731 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.262742 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a27a749-7256-45fc-b79c-4882d5554e9d-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.293354 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.293733 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.333855 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.354184 4969 generic.go:334] "Generic (PLEG): container finished" podID="6a27a749-7256-45fc-b79c-4882d5554e9d" containerID="ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b" exitCode=0 Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.354242 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" event={"ID":"6a27a749-7256-45fc-b79c-4882d5554e9d","Type":"ContainerDied","Data":"ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b"} Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.354301 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" event={"ID":"6a27a749-7256-45fc-b79c-4882d5554e9d","Type":"ContainerDied","Data":"2145e3b8b65ed6b486e1c75f271e3ac34d63e3b86a3a9f4533b9cb50ebc5ca9c"} Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.354321 4969 scope.go:117] "RemoveContainer" containerID="ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.354501 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b8d76ff-7kf7d" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.376832 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pjwv2" event={"ID":"8a0efc12-8421-489a-8fe9-97ae54119cc8","Type":"ContainerStarted","Data":"f0bbfda609e25b7387d00cb20d36147fa2f498a407d4d81dd5a4280917093753"} Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.376881 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.377010 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="612cc256-df44-4bf4-9c25-5a672fc14647" containerName="cinder-scheduler" containerID="cri-o://b4336a0dad20763361d2da0cac58511b89a2ec064aedcf2d2acd8c75b06c7b1d" gracePeriod=30 Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.377901 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="612cc256-df44-4bf4-9c25-5a672fc14647" containerName="probe" containerID="cri-o://e07ee3d3d16b99fa24c649e0da23a789abdff7ab8cfba7afbbcd530ce1ca385a" gracePeriod=30 Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.390682 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.418572 4969 scope.go:117] "RemoveContainer" containerID="4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.424933 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:49 crc kubenswrapper[4969]: W1004 08:35:49.437907 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bbdbff6_e7a2_4269_8039_54f0aa55cfd7.slice/crio-fe44ec4e749c71d696028e1654be799313f7a0f6735916fa2020217833828c27 WatchSource:0}: Error finding container fe44ec4e749c71d696028e1654be799313f7a0f6735916fa2020217833828c27: Status 404 returned error can't find the container with id fe44ec4e749c71d696028e1654be799313f7a0f6735916fa2020217833828c27 Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.456294 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b8d76ff-7kf7d"] Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.469617 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b8d76ff-7kf7d"] Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.498141 4969 scope.go:117] "RemoveContainer" containerID="ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b" Oct 04 08:35:49 crc kubenswrapper[4969]: E1004 08:35:49.500765 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b\": container with ID starting with ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b not found: ID does not exist" containerID="ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.500885 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b"} err="failed to get container status \"ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b\": rpc error: code = NotFound desc = could not find container \"ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b\": container with ID starting with ccefc68410ebf3ed668c2ed05da267388c1b983f392ad8134cf200a50297f13b not found: ID does not exist" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.500955 4969 scope.go:117] "RemoveContainer" containerID="4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056" Oct 04 08:35:49 crc kubenswrapper[4969]: E1004 08:35:49.502233 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056\": container with ID starting with 4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056 not found: ID does not exist" containerID="4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.502260 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056"} err="failed to get container status \"4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056\": rpc error: code = NotFound desc = could not find container \"4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056\": container with ID starting with 4cf1debcae0bc6f0e6cfa5844b8daae9b08d66b8a51a27216fd24235472ab056 not found: ID does not exist" Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.667130 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:35:49 crc kubenswrapper[4969]: I1004 08:35:49.667441 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:35:50 crc kubenswrapper[4969]: I1004 08:35:50.388633 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerStarted","Data":"f8f1477bc08ea62ba746e7e90a1cae22f821ed760a80282a701bf6b88d20a741"} Oct 04 08:35:50 crc kubenswrapper[4969]: I1004 08:35:50.388969 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerStarted","Data":"a3a9eac3dffa52a382b4ea884a154c5a9b1c8d47069e8199ed2497dde6250e6b"} Oct 04 08:35:50 crc kubenswrapper[4969]: I1004 08:35:50.388981 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerStarted","Data":"fe44ec4e749c71d696028e1654be799313f7a0f6735916fa2020217833828c27"} Oct 04 08:35:50 crc kubenswrapper[4969]: I1004 08:35:50.392889 4969 generic.go:334] "Generic (PLEG): container finished" podID="612cc256-df44-4bf4-9c25-5a672fc14647" containerID="e07ee3d3d16b99fa24c649e0da23a789abdff7ab8cfba7afbbcd530ce1ca385a" exitCode=0 Oct 04 08:35:50 crc kubenswrapper[4969]: I1004 08:35:50.392921 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"612cc256-df44-4bf4-9c25-5a672fc14647","Type":"ContainerDied","Data":"e07ee3d3d16b99fa24c649e0da23a789abdff7ab8cfba7afbbcd530ce1ca385a"} Oct 04 08:35:50 crc kubenswrapper[4969]: I1004 08:35:50.395113 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 08:35:51 crc kubenswrapper[4969]: I1004 08:35:51.072482 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a27a749-7256-45fc-b79c-4882d5554e9d" path="/var/lib/kubelet/pods/6a27a749-7256-45fc-b79c-4882d5554e9d/volumes" Oct 04 08:35:51 crc kubenswrapper[4969]: I1004 08:35:51.407898 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:35:51 crc kubenswrapper[4969]: I1004 08:35:51.409181 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerStarted","Data":"1baea3a4c087979ee304755456d7615a32eb59de022d8535c36eac30914e9ef7"} Oct 04 08:35:51 crc kubenswrapper[4969]: I1004 08:35:51.467848 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 08:35:51 crc kubenswrapper[4969]: I1004 08:35:51.468771 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 08:35:52 crc kubenswrapper[4969]: I1004 08:35:52.421972 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerStarted","Data":"64645470dbc46f76b8f27c6b3201be4dd1efb64492a18c295ed6cb227ece5e30"} Oct 04 08:35:52 crc kubenswrapper[4969]: I1004 08:35:52.451664 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.94687331 podStartE2EDuration="4.451641188s" podCreationTimestamp="2025-10-04 08:35:48 +0000 UTC" firstStartedPulling="2025-10-04 08:35:49.497748786 +0000 UTC m=+1177.252017600" lastFinishedPulling="2025-10-04 08:35:52.002516664 +0000 UTC m=+1179.756785478" observedRunningTime="2025-10-04 08:35:52.442628378 +0000 UTC m=+1180.196897192" watchObservedRunningTime="2025-10-04 08:35:52.451641188 +0000 UTC m=+1180.205910002" Oct 04 08:35:52 crc kubenswrapper[4969]: I1004 08:35:52.952809 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:52 crc kubenswrapper[4969]: I1004 08:35:52.952913 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:53 crc kubenswrapper[4969]: I1004 08:35:53.039835 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:53 crc kubenswrapper[4969]: I1004 08:35:53.042039 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:53 crc kubenswrapper[4969]: I1004 08:35:53.436326 4969 generic.go:334] "Generic (PLEG): container finished" podID="612cc256-df44-4bf4-9c25-5a672fc14647" containerID="b4336a0dad20763361d2da0cac58511b89a2ec064aedcf2d2acd8c75b06c7b1d" exitCode=0 Oct 04 08:35:53 crc kubenswrapper[4969]: I1004 08:35:53.437343 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"612cc256-df44-4bf4-9c25-5a672fc14647","Type":"ContainerDied","Data":"b4336a0dad20763361d2da0cac58511b89a2ec064aedcf2d2acd8c75b06c7b1d"} Oct 04 08:35:53 crc kubenswrapper[4969]: I1004 08:35:53.437506 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 08:35:53 crc kubenswrapper[4969]: I1004 08:35:53.437891 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:53 crc kubenswrapper[4969]: I1004 08:35:53.437950 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:54 crc kubenswrapper[4969]: I1004 08:35:54.622514 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 04 08:35:54 crc kubenswrapper[4969]: I1004 08:35:54.622847 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:35:54 crc kubenswrapper[4969]: I1004 08:35:54.623645 4969 scope.go:117] "RemoveContainer" containerID="84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58" Oct 04 08:35:54 crc kubenswrapper[4969]: E1004 08:35:54.623942 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:35:55 crc kubenswrapper[4969]: I1004 08:35:55.493300 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:55 crc kubenswrapper[4969]: I1004 08:35:55.493399 4969 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 08:35:55 crc kubenswrapper[4969]: I1004 08:35:55.645001 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 08:35:57 crc kubenswrapper[4969]: I1004 08:35:57.447339 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:35:57 crc kubenswrapper[4969]: I1004 08:35:57.981718 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:35:57 crc kubenswrapper[4969]: I1004 08:35:57.981955 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="ceilometer-central-agent" containerID="cri-o://a3a9eac3dffa52a382b4ea884a154c5a9b1c8d47069e8199ed2497dde6250e6b" gracePeriod=30 Oct 04 08:35:57 crc kubenswrapper[4969]: I1004 08:35:57.982122 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="proxy-httpd" containerID="cri-o://64645470dbc46f76b8f27c6b3201be4dd1efb64492a18c295ed6cb227ece5e30" gracePeriod=30 Oct 04 08:35:57 crc kubenswrapper[4969]: I1004 08:35:57.982289 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="ceilometer-notification-agent" containerID="cri-o://f8f1477bc08ea62ba746e7e90a1cae22f821ed760a80282a701bf6b88d20a741" gracePeriod=30 Oct 04 08:35:57 crc kubenswrapper[4969]: I1004 08:35:57.982340 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="sg-core" containerID="cri-o://1baea3a4c087979ee304755456d7615a32eb59de022d8535c36eac30914e9ef7" gracePeriod=30 Oct 04 08:35:58 crc kubenswrapper[4969]: I1004 08:35:58.317190 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 08:35:58 crc kubenswrapper[4969]: I1004 08:35:58.523901 4969 generic.go:334] "Generic (PLEG): container finished" podID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerID="64645470dbc46f76b8f27c6b3201be4dd1efb64492a18c295ed6cb227ece5e30" exitCode=0 Oct 04 08:35:58 crc kubenswrapper[4969]: I1004 08:35:58.523941 4969 generic.go:334] "Generic (PLEG): container finished" podID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerID="1baea3a4c087979ee304755456d7615a32eb59de022d8535c36eac30914e9ef7" exitCode=2 Oct 04 08:35:58 crc kubenswrapper[4969]: I1004 08:35:58.523952 4969 generic.go:334] "Generic (PLEG): container finished" podID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerID="a3a9eac3dffa52a382b4ea884a154c5a9b1c8d47069e8199ed2497dde6250e6b" exitCode=0 Oct 04 08:35:58 crc kubenswrapper[4969]: I1004 08:35:58.523995 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerDied","Data":"64645470dbc46f76b8f27c6b3201be4dd1efb64492a18c295ed6cb227ece5e30"} Oct 04 08:35:58 crc kubenswrapper[4969]: I1004 08:35:58.524026 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerDied","Data":"1baea3a4c087979ee304755456d7615a32eb59de022d8535c36eac30914e9ef7"} Oct 04 08:35:58 crc kubenswrapper[4969]: I1004 08:35:58.524062 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerDied","Data":"a3a9eac3dffa52a382b4ea884a154c5a9b1c8d47069e8199ed2497dde6250e6b"} Oct 04 08:35:59 crc kubenswrapper[4969]: I1004 08:35:59.874571 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-78c8455845-gfrwq" Oct 04 08:35:59 crc kubenswrapper[4969]: I1004 08:35:59.952531 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-86d475f9d4-dkrh5"] Oct 04 08:35:59 crc kubenswrapper[4969]: I1004 08:35:59.952929 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-86d475f9d4-dkrh5" podUID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerName="neutron-api" containerID="cri-o://729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915" gracePeriod=30 Oct 04 08:35:59 crc kubenswrapper[4969]: I1004 08:35:59.953297 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-86d475f9d4-dkrh5" podUID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerName="neutron-httpd" containerID="cri-o://b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4" gracePeriod=30 Oct 04 08:36:00 crc kubenswrapper[4969]: I1004 08:36:00.543057 4969 generic.go:334] "Generic (PLEG): container finished" podID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerID="f8f1477bc08ea62ba746e7e90a1cae22f821ed760a80282a701bf6b88d20a741" exitCode=0 Oct 04 08:36:00 crc kubenswrapper[4969]: I1004 08:36:00.543115 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerDied","Data":"f8f1477bc08ea62ba746e7e90a1cae22f821ed760a80282a701bf6b88d20a741"} Oct 04 08:36:00 crc kubenswrapper[4969]: I1004 08:36:00.548401 4969 generic.go:334] "Generic (PLEG): container finished" podID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerID="b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4" exitCode=0 Oct 04 08:36:00 crc kubenswrapper[4969]: I1004 08:36:00.548446 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86d475f9d4-dkrh5" event={"ID":"47ecd91e-5401-418c-a830-6a67c0965cc4","Type":"ContainerDied","Data":"b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4"} Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.056009 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.134132 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79vjk\" (UniqueName: \"kubernetes.io/projected/612cc256-df44-4bf4-9c25-5a672fc14647-kube-api-access-79vjk\") pod \"612cc256-df44-4bf4-9c25-5a672fc14647\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.135036 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-combined-ca-bundle\") pod \"612cc256-df44-4bf4-9c25-5a672fc14647\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.135108 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data\") pod \"612cc256-df44-4bf4-9c25-5a672fc14647\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.135139 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/612cc256-df44-4bf4-9c25-5a672fc14647-etc-machine-id\") pod \"612cc256-df44-4bf4-9c25-5a672fc14647\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.135188 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data-custom\") pod \"612cc256-df44-4bf4-9c25-5a672fc14647\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.135239 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-scripts\") pod \"612cc256-df44-4bf4-9c25-5a672fc14647\" (UID: \"612cc256-df44-4bf4-9c25-5a672fc14647\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.136717 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/612cc256-df44-4bf4-9c25-5a672fc14647-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "612cc256-df44-4bf4-9c25-5a672fc14647" (UID: "612cc256-df44-4bf4-9c25-5a672fc14647"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.141304 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/612cc256-df44-4bf4-9c25-5a672fc14647-kube-api-access-79vjk" (OuterVolumeSpecName: "kube-api-access-79vjk") pod "612cc256-df44-4bf4-9c25-5a672fc14647" (UID: "612cc256-df44-4bf4-9c25-5a672fc14647"). InnerVolumeSpecName "kube-api-access-79vjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.142578 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "612cc256-df44-4bf4-9c25-5a672fc14647" (UID: "612cc256-df44-4bf4-9c25-5a672fc14647"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.143490 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-scripts" (OuterVolumeSpecName: "scripts") pod "612cc256-df44-4bf4-9c25-5a672fc14647" (UID: "612cc256-df44-4bf4-9c25-5a672fc14647"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.225716 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "612cc256-df44-4bf4-9c25-5a672fc14647" (UID: "612cc256-df44-4bf4-9c25-5a672fc14647"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.234089 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.237851 4969 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.237874 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.237885 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79vjk\" (UniqueName: \"kubernetes.io/projected/612cc256-df44-4bf4-9c25-5a672fc14647-kube-api-access-79vjk\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.237895 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.237903 4969 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/612cc256-df44-4bf4-9c25-5a672fc14647-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.340384 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-sg-core-conf-yaml\") pod \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.340473 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-combined-ca-bundle\") pod \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.340608 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4r8h\" (UniqueName: \"kubernetes.io/projected/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-kube-api-access-d4r8h\") pod \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.340714 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-run-httpd\") pod \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.340746 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-config-data\") pod \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.340786 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-scripts\") pod \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.340829 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-log-httpd\") pod \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\" (UID: \"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7\") " Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.341850 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" (UID: "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.343636 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" (UID: "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.346502 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data" (OuterVolumeSpecName: "config-data") pod "612cc256-df44-4bf4-9c25-5a672fc14647" (UID: "612cc256-df44-4bf4-9c25-5a672fc14647"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.348927 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-kube-api-access-d4r8h" (OuterVolumeSpecName: "kube-api-access-d4r8h") pod "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" (UID: "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7"). InnerVolumeSpecName "kube-api-access-d4r8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.349096 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-scripts" (OuterVolumeSpecName: "scripts") pod "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" (UID: "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.377586 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" (UID: "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.442536 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4r8h\" (UniqueName: \"kubernetes.io/projected/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-kube-api-access-d4r8h\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.442564 4969 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.442574 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612cc256-df44-4bf4-9c25-5a672fc14647-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.442582 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.442590 4969 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.442600 4969 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.448547 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-config-data" (OuterVolumeSpecName: "config-data") pod "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" (UID: "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.468124 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" (UID: "8bbdbff6-e7a2-4269-8039-54f0aa55cfd7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.544770 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.544806 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.559136 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bbdbff6-e7a2-4269-8039-54f0aa55cfd7","Type":"ContainerDied","Data":"fe44ec4e749c71d696028e1654be799313f7a0f6735916fa2020217833828c27"} Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.559186 4969 scope.go:117] "RemoveContainer" containerID="64645470dbc46f76b8f27c6b3201be4dd1efb64492a18c295ed6cb227ece5e30" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.559198 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.561035 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"612cc256-df44-4bf4-9c25-5a672fc14647","Type":"ContainerDied","Data":"9b055b417ebaff03faea4f0bd206f01823a7e1031b7eb5568e354b895beda565"} Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.561184 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.566868 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pjwv2" event={"ID":"8a0efc12-8421-489a-8fe9-97ae54119cc8","Type":"ContainerStarted","Data":"8e2e3e952189d7105e232645f1ca31ee92db17c6482688944a3b1999f64fc3fd"} Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.591884 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-pjwv2" podStartSLOduration=2.509543625 podStartE2EDuration="14.591867886s" podCreationTimestamp="2025-10-04 08:35:47 +0000 UTC" firstStartedPulling="2025-10-04 08:35:48.774847158 +0000 UTC m=+1176.529115962" lastFinishedPulling="2025-10-04 08:36:00.857171389 +0000 UTC m=+1188.611440223" observedRunningTime="2025-10-04 08:36:01.582849785 +0000 UTC m=+1189.337118599" watchObservedRunningTime="2025-10-04 08:36:01.591867886 +0000 UTC m=+1189.346136700" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.593865 4969 scope.go:117] "RemoveContainer" containerID="1baea3a4c087979ee304755456d7615a32eb59de022d8535c36eac30914e9ef7" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.624697 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.625660 4969 scope.go:117] "RemoveContainer" containerID="f8f1477bc08ea62ba746e7e90a1cae22f821ed760a80282a701bf6b88d20a741" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.666363 4969 scope.go:117] "RemoveContainer" containerID="a3a9eac3dffa52a382b4ea884a154c5a9b1c8d47069e8199ed2497dde6250e6b" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.682562 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.691775 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.699281 4969 scope.go:117] "RemoveContainer" containerID="e07ee3d3d16b99fa24c649e0da23a789abdff7ab8cfba7afbbcd530ce1ca385a" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706107 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:01 crc kubenswrapper[4969]: E1004 08:36:01.706642 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="ceilometer-central-agent" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706664 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="ceilometer-central-agent" Oct 04 08:36:01 crc kubenswrapper[4969]: E1004 08:36:01.706702 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="612cc256-df44-4bf4-9c25-5a672fc14647" containerName="cinder-scheduler" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706709 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="612cc256-df44-4bf4-9c25-5a672fc14647" containerName="cinder-scheduler" Oct 04 08:36:01 crc kubenswrapper[4969]: E1004 08:36:01.706722 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="sg-core" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706728 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="sg-core" Oct 04 08:36:01 crc kubenswrapper[4969]: E1004 08:36:01.706744 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a27a749-7256-45fc-b79c-4882d5554e9d" containerName="init" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706750 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a27a749-7256-45fc-b79c-4882d5554e9d" containerName="init" Oct 04 08:36:01 crc kubenswrapper[4969]: E1004 08:36:01.706756 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="612cc256-df44-4bf4-9c25-5a672fc14647" containerName="probe" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706761 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="612cc256-df44-4bf4-9c25-5a672fc14647" containerName="probe" Oct 04 08:36:01 crc kubenswrapper[4969]: E1004 08:36:01.706769 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a27a749-7256-45fc-b79c-4882d5554e9d" containerName="dnsmasq-dns" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706774 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a27a749-7256-45fc-b79c-4882d5554e9d" containerName="dnsmasq-dns" Oct 04 08:36:01 crc kubenswrapper[4969]: E1004 08:36:01.706787 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="proxy-httpd" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706794 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="proxy-httpd" Oct 04 08:36:01 crc kubenswrapper[4969]: E1004 08:36:01.706807 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="ceilometer-notification-agent" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706813 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="ceilometer-notification-agent" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706986 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="612cc256-df44-4bf4-9c25-5a672fc14647" containerName="cinder-scheduler" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.706997 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="sg-core" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.707011 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="612cc256-df44-4bf4-9c25-5a672fc14647" containerName="probe" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.707026 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="ceilometer-notification-agent" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.707037 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="ceilometer-central-agent" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.707046 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a27a749-7256-45fc-b79c-4882d5554e9d" containerName="dnsmasq-dns" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.707054 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" containerName="proxy-httpd" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.709289 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.711387 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.713314 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.714672 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.731743 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.750636 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.754033 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.754731 4969 scope.go:117] "RemoveContainer" containerID="b4336a0dad20763361d2da0cac58511b89a2ec064aedcf2d2acd8c75b06c7b1d" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.757137 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.761693 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.862841 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.862888 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l64b\" (UniqueName: \"kubernetes.io/projected/32563c4c-c768-466f-8477-83a180cebc29-kube-api-access-2l64b\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.862924 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-config-data\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.862944 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.862959 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.863006 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-config-data\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.863044 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-scripts\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.863062 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.863077 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7183f88-c473-4e74-8010-f8e7a380b35f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.863232 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrdhr\" (UniqueName: \"kubernetes.io/projected/a7183f88-c473-4e74-8010-f8e7a380b35f-kube-api-access-hrdhr\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.863339 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-log-httpd\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.863377 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-scripts\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.863407 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-run-httpd\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.928604 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:01 crc kubenswrapper[4969]: E1004 08:36:01.929387 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-2l64b log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="32563c4c-c768-466f-8477-83a180cebc29" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.964771 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-scripts\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.965023 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.965103 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7183f88-c473-4e74-8010-f8e7a380b35f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.965174 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7183f88-c473-4e74-8010-f8e7a380b35f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.965263 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrdhr\" (UniqueName: \"kubernetes.io/projected/a7183f88-c473-4e74-8010-f8e7a380b35f-kube-api-access-hrdhr\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.965598 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-log-httpd\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.965722 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-scripts\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.965811 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-run-httpd\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.965890 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.965966 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l64b\" (UniqueName: \"kubernetes.io/projected/32563c4c-c768-466f-8477-83a180cebc29-kube-api-access-2l64b\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.966044 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-config-data\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.966127 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.966202 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.966311 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-config-data\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.966129 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-run-httpd\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.966066 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-log-httpd\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.969918 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-scripts\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.970076 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.970711 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-config-data\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.971070 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.970407 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-config-data\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.974784 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7183f88-c473-4e74-8010-f8e7a380b35f-scripts\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.984871 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l64b\" (UniqueName: \"kubernetes.io/projected/32563c4c-c768-466f-8477-83a180cebc29-kube-api-access-2l64b\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.985370 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrdhr\" (UniqueName: \"kubernetes.io/projected/a7183f88-c473-4e74-8010-f8e7a380b35f-kube-api-access-hrdhr\") pod \"cinder-scheduler-0\" (UID: \"a7183f88-c473-4e74-8010-f8e7a380b35f\") " pod="openstack/cinder-scheduler-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.991677 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:01 crc kubenswrapper[4969]: I1004 08:36:01.992311 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " pod="openstack/ceilometer-0" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.141559 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.577581 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.590061 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.604591 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.679934 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-scripts\") pod \"32563c4c-c768-466f-8477-83a180cebc29\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.680306 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-combined-ca-bundle\") pod \"32563c4c-c768-466f-8477-83a180cebc29\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.680346 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-sg-core-conf-yaml\") pod \"32563c4c-c768-466f-8477-83a180cebc29\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.680401 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-log-httpd\") pod \"32563c4c-c768-466f-8477-83a180cebc29\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.680439 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l64b\" (UniqueName: \"kubernetes.io/projected/32563c4c-c768-466f-8477-83a180cebc29-kube-api-access-2l64b\") pod \"32563c4c-c768-466f-8477-83a180cebc29\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.680471 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-run-httpd\") pod \"32563c4c-c768-466f-8477-83a180cebc29\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.680491 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-config-data\") pod \"32563c4c-c768-466f-8477-83a180cebc29\" (UID: \"32563c4c-c768-466f-8477-83a180cebc29\") " Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.682002 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "32563c4c-c768-466f-8477-83a180cebc29" (UID: "32563c4c-c768-466f-8477-83a180cebc29"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.682515 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "32563c4c-c768-466f-8477-83a180cebc29" (UID: "32563c4c-c768-466f-8477-83a180cebc29"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.686169 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32563c4c-c768-466f-8477-83a180cebc29" (UID: "32563c4c-c768-466f-8477-83a180cebc29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.686789 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-config-data" (OuterVolumeSpecName: "config-data") pod "32563c4c-c768-466f-8477-83a180cebc29" (UID: "32563c4c-c768-466f-8477-83a180cebc29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.686876 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-scripts" (OuterVolumeSpecName: "scripts") pod "32563c4c-c768-466f-8477-83a180cebc29" (UID: "32563c4c-c768-466f-8477-83a180cebc29"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.687027 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32563c4c-c768-466f-8477-83a180cebc29-kube-api-access-2l64b" (OuterVolumeSpecName: "kube-api-access-2l64b") pod "32563c4c-c768-466f-8477-83a180cebc29" (UID: "32563c4c-c768-466f-8477-83a180cebc29"). InnerVolumeSpecName "kube-api-access-2l64b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.690509 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "32563c4c-c768-466f-8477-83a180cebc29" (UID: "32563c4c-c768-466f-8477-83a180cebc29"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.783345 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.783378 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.783389 4969 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.783399 4969 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.783408 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l64b\" (UniqueName: \"kubernetes.io/projected/32563c4c-c768-466f-8477-83a180cebc29-kube-api-access-2l64b\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.783416 4969 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32563c4c-c768-466f-8477-83a180cebc29-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:02 crc kubenswrapper[4969]: I1004 08:36:02.783438 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32563c4c-c768-466f-8477-83a180cebc29-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.066094 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="612cc256-df44-4bf4-9c25-5a672fc14647" path="/var/lib/kubelet/pods/612cc256-df44-4bf4-9c25-5a672fc14647/volumes" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.066848 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bbdbff6-e7a2-4269-8039-54f0aa55cfd7" path="/var/lib/kubelet/pods/8bbdbff6-e7a2-4269-8039-54f0aa55cfd7/volumes" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.589970 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.590107 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a7183f88-c473-4e74-8010-f8e7a380b35f","Type":"ContainerStarted","Data":"7b10f140f368980801e40a0e1efdac86526360f9101ac906b7fa2ebcb517ce1b"} Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.590388 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a7183f88-c473-4e74-8010-f8e7a380b35f","Type":"ContainerStarted","Data":"d16b93a68f0009cbf50f6529a77c058a3211990e683bbaa5dcf57b061082966b"} Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.642792 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.650652 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.664997 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.669608 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.672256 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.710223 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.710353 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.810521 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lz4w\" (UniqueName: \"kubernetes.io/projected/f0021ab7-5417-4f92-a69b-eed20068b066-kube-api-access-8lz4w\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.810669 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-scripts\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.810715 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.810781 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.810850 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-run-httpd\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.811088 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-log-httpd\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.811187 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-config-data\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.912740 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-run-httpd\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.913162 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-log-httpd\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.913197 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-config-data\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.913156 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-run-httpd\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.913274 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lz4w\" (UniqueName: \"kubernetes.io/projected/f0021ab7-5417-4f92-a69b-eed20068b066-kube-api-access-8lz4w\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.913371 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-scripts\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.913408 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.913472 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-log-httpd\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.913498 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.918854 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-scripts\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.920629 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-config-data\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.931051 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.931074 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lz4w\" (UniqueName: \"kubernetes.io/projected/f0021ab7-5417-4f92-a69b-eed20068b066-kube-api-access-8lz4w\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:03 crc kubenswrapper[4969]: I1004 08:36:03.944956 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " pod="openstack/ceilometer-0" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.130858 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.287710 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.425874 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-combined-ca-bundle\") pod \"47ecd91e-5401-418c-a830-6a67c0965cc4\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.427092 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-ovndb-tls-certs\") pod \"47ecd91e-5401-418c-a830-6a67c0965cc4\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.427146 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkjdp\" (UniqueName: \"kubernetes.io/projected/47ecd91e-5401-418c-a830-6a67c0965cc4-kube-api-access-kkjdp\") pod \"47ecd91e-5401-418c-a830-6a67c0965cc4\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.427322 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-config\") pod \"47ecd91e-5401-418c-a830-6a67c0965cc4\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.427386 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-httpd-config\") pod \"47ecd91e-5401-418c-a830-6a67c0965cc4\" (UID: \"47ecd91e-5401-418c-a830-6a67c0965cc4\") " Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.439855 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "47ecd91e-5401-418c-a830-6a67c0965cc4" (UID: "47ecd91e-5401-418c-a830-6a67c0965cc4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.441343 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ecd91e-5401-418c-a830-6a67c0965cc4-kube-api-access-kkjdp" (OuterVolumeSpecName: "kube-api-access-kkjdp") pod "47ecd91e-5401-418c-a830-6a67c0965cc4" (UID: "47ecd91e-5401-418c-a830-6a67c0965cc4"). InnerVolumeSpecName "kube-api-access-kkjdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.478370 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-config" (OuterVolumeSpecName: "config") pod "47ecd91e-5401-418c-a830-6a67c0965cc4" (UID: "47ecd91e-5401-418c-a830-6a67c0965cc4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.480225 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47ecd91e-5401-418c-a830-6a67c0965cc4" (UID: "47ecd91e-5401-418c-a830-6a67c0965cc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.512556 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "47ecd91e-5401-418c-a830-6a67c0965cc4" (UID: "47ecd91e-5401-418c-a830-6a67c0965cc4"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.530244 4969 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.530281 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.530296 4969 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.530311 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkjdp\" (UniqueName: \"kubernetes.io/projected/47ecd91e-5401-418c-a830-6a67c0965cc4-kube-api-access-kkjdp\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.530324 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ecd91e-5401-418c-a830-6a67c0965cc4-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.605330 4969 generic.go:334] "Generic (PLEG): container finished" podID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerID="729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915" exitCode=0 Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.605487 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86d475f9d4-dkrh5" event={"ID":"47ecd91e-5401-418c-a830-6a67c0965cc4","Type":"ContainerDied","Data":"729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915"} Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.605543 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86d475f9d4-dkrh5" event={"ID":"47ecd91e-5401-418c-a830-6a67c0965cc4","Type":"ContainerDied","Data":"c96daad6e33b4ff7512c225598849614639fe61d5de22debd74e31b95208531d"} Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.605579 4969 scope.go:117] "RemoveContainer" containerID="b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.605988 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86d475f9d4-dkrh5" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.617939 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a7183f88-c473-4e74-8010-f8e7a380b35f","Type":"ContainerStarted","Data":"f832725156bcd85ce74265ebf83e03d05bd31c54d16257aea18da7d2c5a1fc08"} Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.649397 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.655608 4969 scope.go:117] "RemoveContainer" containerID="729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.656152 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.656137132 podStartE2EDuration="3.656137132s" podCreationTimestamp="2025-10-04 08:36:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:04.637572899 +0000 UTC m=+1192.391841743" watchObservedRunningTime="2025-10-04 08:36:04.656137132 +0000 UTC m=+1192.410405946" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.680513 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-86d475f9d4-dkrh5"] Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.687998 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-86d475f9d4-dkrh5"] Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.701001 4969 scope.go:117] "RemoveContainer" containerID="b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4" Oct 04 08:36:04 crc kubenswrapper[4969]: E1004 08:36:04.701645 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4\": container with ID starting with b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4 not found: ID does not exist" containerID="b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.701676 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4"} err="failed to get container status \"b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4\": rpc error: code = NotFound desc = could not find container \"b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4\": container with ID starting with b461e782a64f9cda8237c31897e1399f34db495482dbd525c18958b1c81212e4 not found: ID does not exist" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.701699 4969 scope.go:117] "RemoveContainer" containerID="729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915" Oct 04 08:36:04 crc kubenswrapper[4969]: E1004 08:36:04.702114 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915\": container with ID starting with 729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915 not found: ID does not exist" containerID="729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915" Oct 04 08:36:04 crc kubenswrapper[4969]: I1004 08:36:04.702135 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915"} err="failed to get container status \"729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915\": rpc error: code = NotFound desc = could not find container \"729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915\": container with ID starting with 729667120a6fa10b77b9250acbb7b8114cd3f34a2a22d8d4e53d1e1880b4a915 not found: ID does not exist" Oct 04 08:36:05 crc kubenswrapper[4969]: I1004 08:36:05.077079 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32563c4c-c768-466f-8477-83a180cebc29" path="/var/lib/kubelet/pods/32563c4c-c768-466f-8477-83a180cebc29/volumes" Oct 04 08:36:05 crc kubenswrapper[4969]: I1004 08:36:05.078101 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ecd91e-5401-418c-a830-6a67c0965cc4" path="/var/lib/kubelet/pods/47ecd91e-5401-418c-a830-6a67c0965cc4/volumes" Oct 04 08:36:05 crc kubenswrapper[4969]: I1004 08:36:05.633545 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerStarted","Data":"9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec"} Oct 04 08:36:05 crc kubenswrapper[4969]: I1004 08:36:05.633848 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerStarted","Data":"6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130"} Oct 04 08:36:05 crc kubenswrapper[4969]: I1004 08:36:05.633860 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerStarted","Data":"7704c1f08b9307cc7a7945473fcd943c1ae74113aa4e0990ccb9b3519809481d"} Oct 04 08:36:06 crc kubenswrapper[4969]: I1004 08:36:06.055588 4969 scope.go:117] "RemoveContainer" containerID="84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58" Oct 04 08:36:06 crc kubenswrapper[4969]: E1004 08:36:06.056036 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(7144eb34-59d2-4583-8c46-3e56258966da)\"" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" Oct 04 08:36:06 crc kubenswrapper[4969]: I1004 08:36:06.645499 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerStarted","Data":"930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee"} Oct 04 08:36:07 crc kubenswrapper[4969]: I1004 08:36:07.141980 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 08:36:07 crc kubenswrapper[4969]: I1004 08:36:07.658998 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerStarted","Data":"e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506"} Oct 04 08:36:07 crc kubenswrapper[4969]: I1004 08:36:07.659287 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 08:36:07 crc kubenswrapper[4969]: I1004 08:36:07.721594 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.120965259 podStartE2EDuration="4.721575077s" podCreationTimestamp="2025-10-04 08:36:03 +0000 UTC" firstStartedPulling="2025-10-04 08:36:04.674379727 +0000 UTC m=+1192.428648551" lastFinishedPulling="2025-10-04 08:36:07.274989535 +0000 UTC m=+1195.029258369" observedRunningTime="2025-10-04 08:36:07.719124317 +0000 UTC m=+1195.473393131" watchObservedRunningTime="2025-10-04 08:36:07.721575077 +0000 UTC m=+1195.475843891" Oct 04 08:36:12 crc kubenswrapper[4969]: I1004 08:36:12.283284 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 08:36:16 crc kubenswrapper[4969]: I1004 08:36:16.791795 4969 generic.go:334] "Generic (PLEG): container finished" podID="8a0efc12-8421-489a-8fe9-97ae54119cc8" containerID="8e2e3e952189d7105e232645f1ca31ee92db17c6482688944a3b1999f64fc3fd" exitCode=0 Oct 04 08:36:16 crc kubenswrapper[4969]: I1004 08:36:16.791935 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pjwv2" event={"ID":"8a0efc12-8421-489a-8fe9-97ae54119cc8","Type":"ContainerDied","Data":"8e2e3e952189d7105e232645f1ca31ee92db17c6482688944a3b1999f64fc3fd"} Oct 04 08:36:17 crc kubenswrapper[4969]: I1004 08:36:17.056158 4969 scope.go:117] "RemoveContainer" containerID="84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58" Oct 04 08:36:17 crc kubenswrapper[4969]: I1004 08:36:17.805353 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerStarted","Data":"d1921bba8114c64be7e756cc194b0e4d36aae55db8752b7327b3b2167320c331"} Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.319861 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.432042 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-config-data\") pod \"8a0efc12-8421-489a-8fe9-97ae54119cc8\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.432114 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp9s2\" (UniqueName: \"kubernetes.io/projected/8a0efc12-8421-489a-8fe9-97ae54119cc8-kube-api-access-dp9s2\") pod \"8a0efc12-8421-489a-8fe9-97ae54119cc8\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.432225 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-scripts\") pod \"8a0efc12-8421-489a-8fe9-97ae54119cc8\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.432331 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-combined-ca-bundle\") pod \"8a0efc12-8421-489a-8fe9-97ae54119cc8\" (UID: \"8a0efc12-8421-489a-8fe9-97ae54119cc8\") " Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.451693 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a0efc12-8421-489a-8fe9-97ae54119cc8-kube-api-access-dp9s2" (OuterVolumeSpecName: "kube-api-access-dp9s2") pod "8a0efc12-8421-489a-8fe9-97ae54119cc8" (UID: "8a0efc12-8421-489a-8fe9-97ae54119cc8"). InnerVolumeSpecName "kube-api-access-dp9s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.457652 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-scripts" (OuterVolumeSpecName: "scripts") pod "8a0efc12-8421-489a-8fe9-97ae54119cc8" (UID: "8a0efc12-8421-489a-8fe9-97ae54119cc8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.486709 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a0efc12-8421-489a-8fe9-97ae54119cc8" (UID: "8a0efc12-8421-489a-8fe9-97ae54119cc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.512609 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-config-data" (OuterVolumeSpecName: "config-data") pod "8a0efc12-8421-489a-8fe9-97ae54119cc8" (UID: "8a0efc12-8421-489a-8fe9-97ae54119cc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.534924 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.534956 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.534965 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp9s2\" (UniqueName: \"kubernetes.io/projected/8a0efc12-8421-489a-8fe9-97ae54119cc8-kube-api-access-dp9s2\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.534977 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a0efc12-8421-489a-8fe9-97ae54119cc8-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.817550 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pjwv2" event={"ID":"8a0efc12-8421-489a-8fe9-97ae54119cc8","Type":"ContainerDied","Data":"f0bbfda609e25b7387d00cb20d36147fa2f498a407d4d81dd5a4280917093753"} Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.817585 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0bbfda609e25b7387d00cb20d36147fa2f498a407d4d81dd5a4280917093753" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.817649 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pjwv2" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.990532 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 08:36:18 crc kubenswrapper[4969]: E1004 08:36:18.990963 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a0efc12-8421-489a-8fe9-97ae54119cc8" containerName="nova-cell0-conductor-db-sync" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.990978 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a0efc12-8421-489a-8fe9-97ae54119cc8" containerName="nova-cell0-conductor-db-sync" Oct 04 08:36:18 crc kubenswrapper[4969]: E1004 08:36:18.991012 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerName="neutron-api" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.991019 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerName="neutron-api" Oct 04 08:36:18 crc kubenswrapper[4969]: E1004 08:36:18.991042 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerName="neutron-httpd" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.991051 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerName="neutron-httpd" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.991285 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a0efc12-8421-489a-8fe9-97ae54119cc8" containerName="nova-cell0-conductor-db-sync" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.991304 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerName="neutron-httpd" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.991327 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ecd91e-5401-418c-a830-6a67c0965cc4" containerName="neutron-api" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.992242 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.998919 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bczdv" Oct 04 08:36:18 crc kubenswrapper[4969]: I1004 08:36:18.999350 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.013819 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.146618 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60f12325-d22b-4986-a653-854d290f0adc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"60f12325-d22b-4986-a653-854d290f0adc\") " pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.147032 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvrsp\" (UniqueName: \"kubernetes.io/projected/60f12325-d22b-4986-a653-854d290f0adc-kube-api-access-fvrsp\") pod \"nova-cell0-conductor-0\" (UID: \"60f12325-d22b-4986-a653-854d290f0adc\") " pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.147198 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f12325-d22b-4986-a653-854d290f0adc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"60f12325-d22b-4986-a653-854d290f0adc\") " pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.248769 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f12325-d22b-4986-a653-854d290f0adc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"60f12325-d22b-4986-a653-854d290f0adc\") " pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.249077 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60f12325-d22b-4986-a653-854d290f0adc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"60f12325-d22b-4986-a653-854d290f0adc\") " pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.249340 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvrsp\" (UniqueName: \"kubernetes.io/projected/60f12325-d22b-4986-a653-854d290f0adc-kube-api-access-fvrsp\") pod \"nova-cell0-conductor-0\" (UID: \"60f12325-d22b-4986-a653-854d290f0adc\") " pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.252860 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60f12325-d22b-4986-a653-854d290f0adc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"60f12325-d22b-4986-a653-854d290f0adc\") " pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.254644 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f12325-d22b-4986-a653-854d290f0adc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"60f12325-d22b-4986-a653-854d290f0adc\") " pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.289132 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvrsp\" (UniqueName: \"kubernetes.io/projected/60f12325-d22b-4986-a653-854d290f0adc-kube-api-access-fvrsp\") pod \"nova-cell0-conductor-0\" (UID: \"60f12325-d22b-4986-a653-854d290f0adc\") " pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.317844 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.666199 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.666651 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.666716 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.667624 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f952a2138e627189e2388732fb9b24c40ce1c3b15ba69934279b103aa755d5b4"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.667702 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://f952a2138e627189e2388732fb9b24c40ce1c3b15ba69934279b103aa755d5b4" gracePeriod=600 Oct 04 08:36:19 crc kubenswrapper[4969]: W1004 08:36:19.778694 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60f12325_d22b_4986_a653_854d290f0adc.slice/crio-db079a2be45afadee3dbfa165a753bbae4cd0bfe8978e4a0dae0ecf1786da314 WatchSource:0}: Error finding container db079a2be45afadee3dbfa165a753bbae4cd0bfe8978e4a0dae0ecf1786da314: Status 404 returned error can't find the container with id db079a2be45afadee3dbfa165a753bbae4cd0bfe8978e4a0dae0ecf1786da314 Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.780603 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.875146 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="f952a2138e627189e2388732fb9b24c40ce1c3b15ba69934279b103aa755d5b4" exitCode=0 Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.875204 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"f952a2138e627189e2388732fb9b24c40ce1c3b15ba69934279b103aa755d5b4"} Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.876785 4969 scope.go:117] "RemoveContainer" containerID="a0e7ef25f0eb92534c8c59058c5d643ea03dee7ad9ad7534679a9b23c47e914c" Oct 04 08:36:19 crc kubenswrapper[4969]: I1004 08:36:19.878082 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"60f12325-d22b-4986-a653-854d290f0adc","Type":"ContainerStarted","Data":"db079a2be45afadee3dbfa165a753bbae4cd0bfe8978e4a0dae0ecf1786da314"} Oct 04 08:36:20 crc kubenswrapper[4969]: I1004 08:36:20.891213 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"60f12325-d22b-4986-a653-854d290f0adc","Type":"ContainerStarted","Data":"04ef6721bcbc51dd60bf638a06b22506c38debc19c333d4151f3d99ee6297fa6"} Oct 04 08:36:20 crc kubenswrapper[4969]: I1004 08:36:20.891770 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:20 crc kubenswrapper[4969]: I1004 08:36:20.895994 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"139722c0c58580b4c2a5dc3b9a83a0c499c007e51329c9719d5fb5495e97f91c"} Oct 04 08:36:20 crc kubenswrapper[4969]: I1004 08:36:20.932849 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.932832228 podStartE2EDuration="2.932832228s" podCreationTimestamp="2025-10-04 08:36:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:20.91325654 +0000 UTC m=+1208.667525374" watchObservedRunningTime="2025-10-04 08:36:20.932832228 +0000 UTC m=+1208.687101042" Oct 04 08:36:24 crc kubenswrapper[4969]: I1004 08:36:24.623093 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:24 crc kubenswrapper[4969]: I1004 08:36:24.623611 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:24 crc kubenswrapper[4969]: I1004 08:36:24.669632 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:24 crc kubenswrapper[4969]: I1004 08:36:24.994895 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:25 crc kubenswrapper[4969]: I1004 08:36:25.051902 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 04 08:36:26 crc kubenswrapper[4969]: I1004 08:36:26.230475 4969 scope.go:117] "RemoveContainer" containerID="f8c9bda401d4ff35c2e1d5075f74099768d5bcec416d35389e01c82115be5e09" Oct 04 08:36:26 crc kubenswrapper[4969]: I1004 08:36:26.263305 4969 scope.go:117] "RemoveContainer" containerID="3f2a0c9d900fa83797a047787121e6773a927ec043610b8d8e9c2594cf67d237" Oct 04 08:36:26 crc kubenswrapper[4969]: I1004 08:36:26.295468 4969 scope.go:117] "RemoveContainer" containerID="b0cddc45a729fbee29630d542e4f295b87b2ff70a3f2bb09831b96d871c3d669" Oct 04 08:36:26 crc kubenswrapper[4969]: I1004 08:36:26.979789 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" containerID="cri-o://d1921bba8114c64be7e756cc194b0e4d36aae55db8752b7327b3b2167320c331" gracePeriod=30 Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.365370 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.862134 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-9hbl9"] Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.863324 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.865766 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.866129 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.874388 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-scripts\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.874780 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.874804 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-config-data\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.874883 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j72nj\" (UniqueName: \"kubernetes.io/projected/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-kube-api-access-j72nj\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.888267 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-9hbl9"] Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.976308 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j72nj\" (UniqueName: \"kubernetes.io/projected/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-kube-api-access-j72nj\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.976457 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-scripts\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.976490 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.976508 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-config-data\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.982750 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-config-data\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:29 crc kubenswrapper[4969]: I1004 08:36:29.986167 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:29.995902 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-scripts\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.015890 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j72nj\" (UniqueName: \"kubernetes.io/projected/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-kube-api-access-j72nj\") pod \"nova-cell0-cell-mapping-9hbl9\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.076088 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.078183 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.082976 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.112582 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.185737 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-logs\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.185770 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.185812 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhrc6\" (UniqueName: \"kubernetes.io/projected/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-kube-api-access-nhrc6\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.185861 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-config-data\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.226843 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.240702 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.242543 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.247201 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.257735 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6759cf46f9-qjbd6"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.259197 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.273266 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287509 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-config-data\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287544 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-svc\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287573 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287594 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589f782e-3da1-4dd5-b652-672955ea4bbc-logs\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287620 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-swift-storage-0\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287655 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-logs\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287675 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287700 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhrc6\" (UniqueName: \"kubernetes.io/projected/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-kube-api-access-nhrc6\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287719 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-nb\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287739 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-sb\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287775 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txwz2\" (UniqueName: \"kubernetes.io/projected/589f782e-3da1-4dd5-b652-672955ea4bbc-kube-api-access-txwz2\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287794 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-config-data\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287810 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d8db\" (UniqueName: \"kubernetes.io/projected/f764da96-9012-4a0a-9144-4922906ed2d3-kube-api-access-4d8db\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.287831 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-config\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.288263 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-logs\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.315269 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-config-data\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.324341 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.328396 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6759cf46f9-qjbd6"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.357098 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhrc6\" (UniqueName: \"kubernetes.io/projected/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-kube-api-access-nhrc6\") pod \"nova-metadata-0\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.396489 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.397895 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.404750 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.409766 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.410625 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.410633 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-nb\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.410722 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-sb\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.410811 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txwz2\" (UniqueName: \"kubernetes.io/projected/589f782e-3da1-4dd5-b652-672955ea4bbc-kube-api-access-txwz2\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.410841 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d8db\" (UniqueName: \"kubernetes.io/projected/f764da96-9012-4a0a-9144-4922906ed2d3-kube-api-access-4d8db\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.410880 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-config\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.411029 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-config-data\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.411043 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-svc\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.411086 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.411127 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589f782e-3da1-4dd5-b652-672955ea4bbc-logs\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.411170 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-swift-storage-0\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.424761 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589f782e-3da1-4dd5-b652-672955ea4bbc-logs\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.431347 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-config-data\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.439264 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-nb\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.442177 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-svc\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.451134 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-config\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.454240 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-sb\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.456027 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-swift-storage-0\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.459056 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.489143 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txwz2\" (UniqueName: \"kubernetes.io/projected/589f782e-3da1-4dd5-b652-672955ea4bbc-kube-api-access-txwz2\") pod \"nova-api-0\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.491003 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d8db\" (UniqueName: \"kubernetes.io/projected/f764da96-9012-4a0a-9144-4922906ed2d3-kube-api-access-4d8db\") pod \"dnsmasq-dns-6759cf46f9-qjbd6\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.514614 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnlmd\" (UniqueName: \"kubernetes.io/projected/6dfd83e4-c050-45a4-905d-331a3aff0f9e-kube-api-access-vnlmd\") pod \"nova-cell1-novncproxy-0\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.514693 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.514752 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.573165 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.574498 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.575941 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.581925 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.582265 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.616543 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.617480 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjrdk\" (UniqueName: \"kubernetes.io/projected/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-kube-api-access-zjrdk\") pod \"nova-scheduler-0\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.617532 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnlmd\" (UniqueName: \"kubernetes.io/projected/6dfd83e4-c050-45a4-905d-331a3aff0f9e-kube-api-access-vnlmd\") pod \"nova-cell1-novncproxy-0\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.617580 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-config-data\") pod \"nova-scheduler-0\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.617599 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.617666 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.617714 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.625454 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.629330 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.658806 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnlmd\" (UniqueName: \"kubernetes.io/projected/6dfd83e4-c050-45a4-905d-331a3aff0f9e-kube-api-access-vnlmd\") pod \"nova-cell1-novncproxy-0\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.719858 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjrdk\" (UniqueName: \"kubernetes.io/projected/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-kube-api-access-zjrdk\") pod \"nova-scheduler-0\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.719983 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-config-data\") pod \"nova-scheduler-0\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.720079 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.729956 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-config-data\") pod \"nova-scheduler-0\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.732600 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.756653 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjrdk\" (UniqueName: \"kubernetes.io/projected/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-kube-api-access-zjrdk\") pod \"nova-scheduler-0\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.946149 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:30 crc kubenswrapper[4969]: I1004 08:36:30.963247 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.037984 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-9hbl9"] Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.212405 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.232525 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9xhfr"] Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.233889 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.237028 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.248476 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.254636 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9xhfr"] Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.302967 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.324762 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6759cf46f9-qjbd6"] Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.341197 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb92c\" (UniqueName: \"kubernetes.io/projected/2160e9f1-8d10-4ddf-931f-972f80fed6ba-kube-api-access-vb92c\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.341247 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-config-data\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.341416 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-scripts\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.341488 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.442755 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-scripts\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.442792 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.442900 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb92c\" (UniqueName: \"kubernetes.io/projected/2160e9f1-8d10-4ddf-931f-972f80fed6ba-kube-api-access-vb92c\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.442926 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-config-data\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.447173 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-scripts\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.454379 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.460875 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-config-data\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.462187 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb92c\" (UniqueName: \"kubernetes.io/projected/2160e9f1-8d10-4ddf-931f-972f80fed6ba-kube-api-access-vb92c\") pod \"nova-cell1-conductor-db-sync-9xhfr\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.568392 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.581750 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:36:31 crc kubenswrapper[4969]: I1004 08:36:31.609413 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:31 crc kubenswrapper[4969]: W1004 08:36:31.615233 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dfd83e4_c050_45a4_905d_331a3aff0f9e.slice/crio-0fdaad971a78db81e70f465a5152b637eec9933d182ed9ebe6b2b22242ceb471 WatchSource:0}: Error finding container 0fdaad971a78db81e70f465a5152b637eec9933d182ed9ebe6b2b22242ceb471: Status 404 returned error can't find the container with id 0fdaad971a78db81e70f465a5152b637eec9933d182ed9ebe6b2b22242ceb471 Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.083654 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9hbl9" event={"ID":"cebc3b1f-5a32-4255-ae14-4fea09ad9f12","Type":"ContainerStarted","Data":"5c3ecbc26a9d98cfe05688d490ec149ed5852e06bd1bf98f08311d10f1aef34a"} Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.084223 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9hbl9" event={"ID":"cebc3b1f-5a32-4255-ae14-4fea09ad9f12","Type":"ContainerStarted","Data":"5fadb0a2b6a46c181240deea19ccf779527b5d67d1559c0407192eed8c75e94c"} Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.090052 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7","Type":"ContainerStarted","Data":"2a0ee03f3a0c6c58ed41cf53da1f45d00e9bc4b099611dc15fd8e0307cca942b"} Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.101640 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-9hbl9" podStartSLOduration=3.101618628 podStartE2EDuration="3.101618628s" podCreationTimestamp="2025-10-04 08:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:32.097224611 +0000 UTC m=+1219.851493415" watchObservedRunningTime="2025-10-04 08:36:32.101618628 +0000 UTC m=+1219.855887442" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.102346 4969 generic.go:334] "Generic (PLEG): container finished" podID="7144eb34-59d2-4583-8c46-3e56258966da" containerID="d1921bba8114c64be7e756cc194b0e4d36aae55db8752b7327b3b2167320c331" exitCode=0 Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.102434 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerDied","Data":"d1921bba8114c64be7e756cc194b0e4d36aae55db8752b7327b3b2167320c331"} Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.102470 4969 scope.go:117] "RemoveContainer" containerID="84f7394abfb8c6cfd82c58ca174ae099a7777fb0e598193350041e6a45903f58" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.107310 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe","Type":"ContainerStarted","Data":"da0a739eb7373234ae079dfe2a2a6ddc3b4cb6905b5a5ce78684652a0b0c5779"} Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.112025 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6dfd83e4-c050-45a4-905d-331a3aff0f9e","Type":"ContainerStarted","Data":"0fdaad971a78db81e70f465a5152b637eec9933d182ed9ebe6b2b22242ceb471"} Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.124373 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589f782e-3da1-4dd5-b652-672955ea4bbc","Type":"ContainerStarted","Data":"e049c26d48397831bd3ecef749e541966b8ef6ea1efb4df9a3aca8c674e2c6f8"} Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.125543 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9xhfr"] Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.129789 4969 generic.go:334] "Generic (PLEG): container finished" podID="f764da96-9012-4a0a-9144-4922906ed2d3" containerID="f94b523751c4b5bba921ee496cbdac61cd6117da890ca466756822e6ced2f46c" exitCode=0 Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.129819 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" event={"ID":"f764da96-9012-4a0a-9144-4922906ed2d3","Type":"ContainerDied","Data":"f94b523751c4b5bba921ee496cbdac61cd6117da890ca466756822e6ced2f46c"} Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.129837 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" event={"ID":"f764da96-9012-4a0a-9144-4922906ed2d3","Type":"ContainerStarted","Data":"4350deba044ad72445a474dae4c995b511dfb8e2d4a0b7f1384d723bb498e963"} Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.365571 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.489163 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-custom-prometheus-ca\") pod \"7144eb34-59d2-4583-8c46-3e56258966da\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.489275 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-combined-ca-bundle\") pod \"7144eb34-59d2-4583-8c46-3e56258966da\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.489318 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-config-data\") pod \"7144eb34-59d2-4583-8c46-3e56258966da\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.489410 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7144eb34-59d2-4583-8c46-3e56258966da-logs\") pod \"7144eb34-59d2-4583-8c46-3e56258966da\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.489481 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw6gk\" (UniqueName: \"kubernetes.io/projected/7144eb34-59d2-4583-8c46-3e56258966da-kube-api-access-mw6gk\") pod \"7144eb34-59d2-4583-8c46-3e56258966da\" (UID: \"7144eb34-59d2-4583-8c46-3e56258966da\") " Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.492918 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7144eb34-59d2-4583-8c46-3e56258966da-logs" (OuterVolumeSpecName: "logs") pod "7144eb34-59d2-4583-8c46-3e56258966da" (UID: "7144eb34-59d2-4583-8c46-3e56258966da"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.496018 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7144eb34-59d2-4583-8c46-3e56258966da-kube-api-access-mw6gk" (OuterVolumeSpecName: "kube-api-access-mw6gk") pod "7144eb34-59d2-4583-8c46-3e56258966da" (UID: "7144eb34-59d2-4583-8c46-3e56258966da"). InnerVolumeSpecName "kube-api-access-mw6gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.515327 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "7144eb34-59d2-4583-8c46-3e56258966da" (UID: "7144eb34-59d2-4583-8c46-3e56258966da"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.565082 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7144eb34-59d2-4583-8c46-3e56258966da" (UID: "7144eb34-59d2-4583-8c46-3e56258966da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.594620 4969 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.594648 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.594659 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7144eb34-59d2-4583-8c46-3e56258966da-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.594669 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw6gk\" (UniqueName: \"kubernetes.io/projected/7144eb34-59d2-4583-8c46-3e56258966da-kube-api-access-mw6gk\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.647664 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-config-data" (OuterVolumeSpecName: "config-data") pod "7144eb34-59d2-4583-8c46-3e56258966da" (UID: "7144eb34-59d2-4583-8c46-3e56258966da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:32 crc kubenswrapper[4969]: I1004 08:36:32.698571 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7144eb34-59d2-4583-8c46-3e56258966da-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.199837 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" event={"ID":"f764da96-9012-4a0a-9144-4922906ed2d3","Type":"ContainerStarted","Data":"7fc94259ee054c59317135db33a8f1abdbf881dd61b963611392daed882ea91b"} Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.200101 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.210457 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9xhfr" event={"ID":"2160e9f1-8d10-4ddf-931f-972f80fed6ba","Type":"ContainerStarted","Data":"906c11d6be4da220c5de9e0d3c69f0a215ef98e2509527008b4d4ab455a32467"} Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.210509 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9xhfr" event={"ID":"2160e9f1-8d10-4ddf-931f-972f80fed6ba","Type":"ContainerStarted","Data":"92e5c94264bec744e06c0bd7cfa1ccfd83d213a99816a217c7ee5e9a1bc2300c"} Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.217211 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"7144eb34-59d2-4583-8c46-3e56258966da","Type":"ContainerDied","Data":"fdd644be47a44d016c267448c3c9ead683598afd4ce87d199bde465d7e7a6b88"} Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.217282 4969 scope.go:117] "RemoveContainer" containerID="d1921bba8114c64be7e756cc194b0e4d36aae55db8752b7327b3b2167320c331" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.217458 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.233326 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-9xhfr" podStartSLOduration=2.233294706 podStartE2EDuration="2.233294706s" podCreationTimestamp="2025-10-04 08:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:33.231552773 +0000 UTC m=+1220.985821577" watchObservedRunningTime="2025-10-04 08:36:33.233294706 +0000 UTC m=+1220.987563520" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.270933 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" podStartSLOduration=3.270915694 podStartE2EDuration="3.270915694s" podCreationTimestamp="2025-10-04 08:36:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:33.256870301 +0000 UTC m=+1221.011139115" watchObservedRunningTime="2025-10-04 08:36:33.270915694 +0000 UTC m=+1221.025184508" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.298390 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.317803 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.344151 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 04 08:36:33 crc kubenswrapper[4969]: E1004 08:36:33.345306 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.345326 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: E1004 08:36:33.345340 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.345347 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: E1004 08:36:33.345383 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.345389 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: E1004 08:36:33.345407 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.345415 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.347930 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.347955 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.348109 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.349648 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.355050 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.357561 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.432319 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f21dc45-2a75-427d-a2c1-632c0fa0428c-config-data\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.432391 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f21dc45-2a75-427d-a2c1-632c0fa0428c-logs\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.432431 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6f21dc45-2a75-427d-a2c1-632c0fa0428c-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.432506 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f21dc45-2a75-427d-a2c1-632c0fa0428c-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.432719 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f48k\" (UniqueName: \"kubernetes.io/projected/6f21dc45-2a75-427d-a2c1-632c0fa0428c-kube-api-access-7f48k\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.534158 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f21dc45-2a75-427d-a2c1-632c0fa0428c-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.534415 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f48k\" (UniqueName: \"kubernetes.io/projected/6f21dc45-2a75-427d-a2c1-632c0fa0428c-kube-api-access-7f48k\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.534494 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f21dc45-2a75-427d-a2c1-632c0fa0428c-config-data\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.534537 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f21dc45-2a75-427d-a2c1-632c0fa0428c-logs\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.534556 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6f21dc45-2a75-427d-a2c1-632c0fa0428c-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.535324 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f21dc45-2a75-427d-a2c1-632c0fa0428c-logs\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.539597 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f21dc45-2a75-427d-a2c1-632c0fa0428c-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.539940 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6f21dc45-2a75-427d-a2c1-632c0fa0428c-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.548094 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f21dc45-2a75-427d-a2c1-632c0fa0428c-config-data\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.551797 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f48k\" (UniqueName: \"kubernetes.io/projected/6f21dc45-2a75-427d-a2c1-632c0fa0428c-kube-api-access-7f48k\") pod \"watcher-decision-engine-0\" (UID: \"6f21dc45-2a75-427d-a2c1-632c0fa0428c\") " pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.679784 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.853007 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:33 crc kubenswrapper[4969]: I1004 08:36:33.905591 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 08:36:34 crc kubenswrapper[4969]: I1004 08:36:34.141254 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 08:36:35 crc kubenswrapper[4969]: I1004 08:36:35.067670 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7144eb34-59d2-4583-8c46-3e56258966da" path="/var/lib/kubelet/pods/7144eb34-59d2-4583-8c46-3e56258966da/volumes" Oct 04 08:36:35 crc kubenswrapper[4969]: I1004 08:36:35.942049 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 04 08:36:35 crc kubenswrapper[4969]: W1004 08:36:35.942311 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f21dc45_2a75_427d_a2c1_632c0fa0428c.slice/crio-a46305ec9ede2acbd122df1371ba015f87619266def1d085573d54b70fae8014 WatchSource:0}: Error finding container a46305ec9ede2acbd122df1371ba015f87619266def1d085573d54b70fae8014: Status 404 returned error can't find the container with id a46305ec9ede2acbd122df1371ba015f87619266def1d085573d54b70fae8014 Oct 04 08:36:36 crc kubenswrapper[4969]: I1004 08:36:36.252109 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7","Type":"ContainerStarted","Data":"4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047"} Oct 04 08:36:36 crc kubenswrapper[4969]: I1004 08:36:36.254303 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6f21dc45-2a75-427d-a2c1-632c0fa0428c","Type":"ContainerStarted","Data":"a46305ec9ede2acbd122df1371ba015f87619266def1d085573d54b70fae8014"} Oct 04 08:36:36 crc kubenswrapper[4969]: I1004 08:36:36.255894 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe","Type":"ContainerStarted","Data":"9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7"} Oct 04 08:36:36 crc kubenswrapper[4969]: I1004 08:36:36.257377 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6dfd83e4-c050-45a4-905d-331a3aff0f9e","Type":"ContainerStarted","Data":"8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72"} Oct 04 08:36:36 crc kubenswrapper[4969]: I1004 08:36:36.257477 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6dfd83e4-c050-45a4-905d-331a3aff0f9e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72" gracePeriod=30 Oct 04 08:36:36 crc kubenswrapper[4969]: I1004 08:36:36.259321 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589f782e-3da1-4dd5-b652-672955ea4bbc","Type":"ContainerStarted","Data":"f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff"} Oct 04 08:36:36 crc kubenswrapper[4969]: I1004 08:36:36.274566 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.488883833 podStartE2EDuration="6.274544251s" podCreationTimestamp="2025-10-04 08:36:30 +0000 UTC" firstStartedPulling="2025-10-04 08:36:31.628063988 +0000 UTC m=+1219.382332802" lastFinishedPulling="2025-10-04 08:36:35.413724406 +0000 UTC m=+1223.167993220" observedRunningTime="2025-10-04 08:36:36.272354927 +0000 UTC m=+1224.026623751" watchObservedRunningTime="2025-10-04 08:36:36.274544251 +0000 UTC m=+1224.028813065" Oct 04 08:36:36 crc kubenswrapper[4969]: I1004 08:36:36.299394 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5030810580000002 podStartE2EDuration="6.299373596s" podCreationTimestamp="2025-10-04 08:36:30 +0000 UTC" firstStartedPulling="2025-10-04 08:36:31.62857355 +0000 UTC m=+1219.382842364" lastFinishedPulling="2025-10-04 08:36:35.424866048 +0000 UTC m=+1223.179134902" observedRunningTime="2025-10-04 08:36:36.288140462 +0000 UTC m=+1224.042409286" watchObservedRunningTime="2025-10-04 08:36:36.299373596 +0000 UTC m=+1224.053642410" Oct 04 08:36:37 crc kubenswrapper[4969]: I1004 08:36:37.268603 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589f782e-3da1-4dd5-b652-672955ea4bbc","Type":"ContainerStarted","Data":"b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730"} Oct 04 08:36:37 crc kubenswrapper[4969]: I1004 08:36:37.271780 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7","Type":"ContainerStarted","Data":"41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29"} Oct 04 08:36:37 crc kubenswrapper[4969]: I1004 08:36:37.271859 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerName="nova-metadata-log" containerID="cri-o://4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047" gracePeriod=30 Oct 04 08:36:37 crc kubenswrapper[4969]: I1004 08:36:37.271949 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerName="nova-metadata-metadata" containerID="cri-o://41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29" gracePeriod=30 Oct 04 08:36:37 crc kubenswrapper[4969]: I1004 08:36:37.273960 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6f21dc45-2a75-427d-a2c1-632c0fa0428c","Type":"ContainerStarted","Data":"a487a964c31d995a5edea4e2b9d4fd0022cb7b92a756f7d3821b1d36c6a368d3"} Oct 04 08:36:37 crc kubenswrapper[4969]: I1004 08:36:37.289791 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.155297732 podStartE2EDuration="7.289774155s" podCreationTimestamp="2025-10-04 08:36:30 +0000 UTC" firstStartedPulling="2025-10-04 08:36:31.282094412 +0000 UTC m=+1219.036363226" lastFinishedPulling="2025-10-04 08:36:35.416570845 +0000 UTC m=+1223.170839649" observedRunningTime="2025-10-04 08:36:37.285747967 +0000 UTC m=+1225.040016781" watchObservedRunningTime="2025-10-04 08:36:37.289774155 +0000 UTC m=+1225.044042969" Oct 04 08:36:37 crc kubenswrapper[4969]: I1004 08:36:37.306478 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=4.306459743 podStartE2EDuration="4.306459743s" podCreationTimestamp="2025-10-04 08:36:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:37.302398783 +0000 UTC m=+1225.056667597" watchObservedRunningTime="2025-10-04 08:36:37.306459743 +0000 UTC m=+1225.060728557" Oct 04 08:36:37 crc kubenswrapper[4969]: I1004 08:36:37.325333 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.167940279 podStartE2EDuration="7.325311712s" podCreationTimestamp="2025-10-04 08:36:30 +0000 UTC" firstStartedPulling="2025-10-04 08:36:31.257385498 +0000 UTC m=+1219.011654312" lastFinishedPulling="2025-10-04 08:36:35.414756931 +0000 UTC m=+1223.169025745" observedRunningTime="2025-10-04 08:36:37.321394116 +0000 UTC m=+1225.075662930" watchObservedRunningTime="2025-10-04 08:36:37.325311712 +0000 UTC m=+1225.079580526" Oct 04 08:36:37 crc kubenswrapper[4969]: I1004 08:36:37.930136 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.058166 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-logs\") pod \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.058383 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhrc6\" (UniqueName: \"kubernetes.io/projected/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-kube-api-access-nhrc6\") pod \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.058561 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-combined-ca-bundle\") pod \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.058613 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-config-data\") pod \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\" (UID: \"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7\") " Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.060650 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-logs" (OuterVolumeSpecName: "logs") pod "a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" (UID: "a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.098198 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-kube-api-access-nhrc6" (OuterVolumeSpecName: "kube-api-access-nhrc6") pod "a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" (UID: "a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7"). InnerVolumeSpecName "kube-api-access-nhrc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.110077 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-config-data" (OuterVolumeSpecName: "config-data") pod "a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" (UID: "a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.119623 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" (UID: "a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.161656 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhrc6\" (UniqueName: \"kubernetes.io/projected/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-kube-api-access-nhrc6\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.161688 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.161697 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.161707 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.290027 4969 generic.go:334] "Generic (PLEG): container finished" podID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerID="41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29" exitCode=0 Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.290063 4969 generic.go:334] "Generic (PLEG): container finished" podID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerID="4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047" exitCode=143 Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.290080 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.290108 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7","Type":"ContainerDied","Data":"41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29"} Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.290176 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7","Type":"ContainerDied","Data":"4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047"} Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.290187 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7","Type":"ContainerDied","Data":"2a0ee03f3a0c6c58ed41cf53da1f45d00e9bc4b099611dc15fd8e0307cca942b"} Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.290202 4969 scope.go:117] "RemoveContainer" containerID="41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.326488 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.339562 4969 scope.go:117] "RemoveContainer" containerID="4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.339680 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.368486 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:38 crc kubenswrapper[4969]: E1004 08:36:38.368939 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.368957 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:38 crc kubenswrapper[4969]: E1004 08:36:38.368978 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerName="nova-metadata-log" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.368986 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerName="nova-metadata-log" Oct 04 08:36:38 crc kubenswrapper[4969]: E1004 08:36:38.369003 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerName="nova-metadata-metadata" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.369011 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerName="nova-metadata-metadata" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.369231 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerName="nova-metadata-log" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.369260 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.369270 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="7144eb34-59d2-4583-8c46-3e56258966da" containerName="watcher-decision-engine" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.369279 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" containerName="nova-metadata-metadata" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.370538 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.374647 4969 scope.go:117] "RemoveContainer" containerID="41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29" Oct 04 08:36:38 crc kubenswrapper[4969]: E1004 08:36:38.378569 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29\": container with ID starting with 41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29 not found: ID does not exist" containerID="41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.378626 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29"} err="failed to get container status \"41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29\": rpc error: code = NotFound desc = could not find container \"41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29\": container with ID starting with 41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29 not found: ID does not exist" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.378653 4969 scope.go:117] "RemoveContainer" containerID="4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047" Oct 04 08:36:38 crc kubenswrapper[4969]: E1004 08:36:38.378959 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047\": container with ID starting with 4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047 not found: ID does not exist" containerID="4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.379000 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047"} err="failed to get container status \"4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047\": rpc error: code = NotFound desc = could not find container \"4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047\": container with ID starting with 4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047 not found: ID does not exist" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.379027 4969 scope.go:117] "RemoveContainer" containerID="41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.379541 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29"} err="failed to get container status \"41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29\": rpc error: code = NotFound desc = could not find container \"41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29\": container with ID starting with 41741a8c8f4e63740cca9e024ab06f235eec6d616e4e636aeadbe3579f2e3c29 not found: ID does not exist" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.379562 4969 scope.go:117] "RemoveContainer" containerID="4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.379832 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.380247 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047"} err="failed to get container status \"4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047\": rpc error: code = NotFound desc = could not find container \"4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047\": container with ID starting with 4d27555bea97d2434e20b83a317e1f45b05f133469c77d3e4202aa4b5dfc2047 not found: ID does not exist" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.380303 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.383734 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.467177 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-config-data\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.467265 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-logs\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.467316 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.467347 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjmn5\" (UniqueName: \"kubernetes.io/projected/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-kube-api-access-mjmn5\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.467440 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.569249 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.569776 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-config-data\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.569900 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-logs\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.570009 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.570095 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjmn5\" (UniqueName: \"kubernetes.io/projected/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-kube-api-access-mjmn5\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.570382 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-logs\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.573914 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.574277 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.585558 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-config-data\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.586129 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjmn5\" (UniqueName: \"kubernetes.io/projected/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-kube-api-access-mjmn5\") pod \"nova-metadata-0\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.718608 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.887759 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 08:36:38 crc kubenswrapper[4969]: I1004 08:36:38.887973 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="0e92558f-4f16-443e-b9a7-1c7f76013d2f" containerName="kube-state-metrics" containerID="cri-o://4d344506dabb32d7f6d8cab02f1a6d2d7772fd46272eab42fc4d4d7c1b3567ac" gracePeriod=30 Oct 04 08:36:39 crc kubenswrapper[4969]: I1004 08:36:39.069478 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7" path="/var/lib/kubelet/pods/a9dc07cb-efe5-4e40-b5ba-1184b7dfc9e7/volumes" Oct 04 08:36:39 crc kubenswrapper[4969]: I1004 08:36:39.211212 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:39 crc kubenswrapper[4969]: W1004 08:36:39.218059 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91ad8cc9_ec8b_4a24_a98f_036ec933a7b2.slice/crio-a1e91b370b353ed4091f886237a6f9bb30927dbf531a6e98cd60c975ca5d0630 WatchSource:0}: Error finding container a1e91b370b353ed4091f886237a6f9bb30927dbf531a6e98cd60c975ca5d0630: Status 404 returned error can't find the container with id a1e91b370b353ed4091f886237a6f9bb30927dbf531a6e98cd60c975ca5d0630 Oct 04 08:36:39 crc kubenswrapper[4969]: I1004 08:36:39.304777 4969 generic.go:334] "Generic (PLEG): container finished" podID="0e92558f-4f16-443e-b9a7-1c7f76013d2f" containerID="4d344506dabb32d7f6d8cab02f1a6d2d7772fd46272eab42fc4d4d7c1b3567ac" exitCode=2 Oct 04 08:36:39 crc kubenswrapper[4969]: I1004 08:36:39.304835 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0e92558f-4f16-443e-b9a7-1c7f76013d2f","Type":"ContainerDied","Data":"4d344506dabb32d7f6d8cab02f1a6d2d7772fd46272eab42fc4d4d7c1b3567ac"} Oct 04 08:36:39 crc kubenswrapper[4969]: I1004 08:36:39.307144 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2","Type":"ContainerStarted","Data":"a1e91b370b353ed4091f886237a6f9bb30927dbf531a6e98cd60c975ca5d0630"} Oct 04 08:36:39 crc kubenswrapper[4969]: I1004 08:36:39.338866 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 08:36:39 crc kubenswrapper[4969]: I1004 08:36:39.395335 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbdbm\" (UniqueName: \"kubernetes.io/projected/0e92558f-4f16-443e-b9a7-1c7f76013d2f-kube-api-access-pbdbm\") pod \"0e92558f-4f16-443e-b9a7-1c7f76013d2f\" (UID: \"0e92558f-4f16-443e-b9a7-1c7f76013d2f\") " Oct 04 08:36:39 crc kubenswrapper[4969]: I1004 08:36:39.431646 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e92558f-4f16-443e-b9a7-1c7f76013d2f-kube-api-access-pbdbm" (OuterVolumeSpecName: "kube-api-access-pbdbm") pod "0e92558f-4f16-443e-b9a7-1c7f76013d2f" (UID: "0e92558f-4f16-443e-b9a7-1c7f76013d2f"). InnerVolumeSpecName "kube-api-access-pbdbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:39 crc kubenswrapper[4969]: I1004 08:36:39.497777 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbdbm\" (UniqueName: \"kubernetes.io/projected/0e92558f-4f16-443e-b9a7-1c7f76013d2f-kube-api-access-pbdbm\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.318631 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0e92558f-4f16-443e-b9a7-1c7f76013d2f","Type":"ContainerDied","Data":"97ca7e0f68581a2d2c39116e1db954233259dad3244f419260935f709390b9ff"} Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.318975 4969 scope.go:117] "RemoveContainer" containerID="4d344506dabb32d7f6d8cab02f1a6d2d7772fd46272eab42fc4d4d7c1b3567ac" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.318662 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.326604 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2","Type":"ContainerStarted","Data":"074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570"} Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.326647 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2","Type":"ContainerStarted","Data":"533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811"} Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.353939 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.3539215589999998 podStartE2EDuration="2.353921559s" podCreationTimestamp="2025-10-04 08:36:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:40.348747013 +0000 UTC m=+1228.103015837" watchObservedRunningTime="2025-10-04 08:36:40.353921559 +0000 UTC m=+1228.108190363" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.380796 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.398964 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.411391 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 08:36:40 crc kubenswrapper[4969]: E1004 08:36:40.411974 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e92558f-4f16-443e-b9a7-1c7f76013d2f" containerName="kube-state-metrics" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.411996 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e92558f-4f16-443e-b9a7-1c7f76013d2f" containerName="kube-state-metrics" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.412261 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e92558f-4f16-443e-b9a7-1c7f76013d2f" containerName="kube-state-metrics" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.413171 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.414956 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.416048 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.425768 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.515340 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5385fdd1-eb06-4d69-a383-236a55505108-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.515663 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d6wl\" (UniqueName: \"kubernetes.io/projected/5385fdd1-eb06-4d69-a383-236a55505108-kube-api-access-2d6wl\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.515687 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5385fdd1-eb06-4d69-a383-236a55505108-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.516012 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5385fdd1-eb06-4d69-a383-236a55505108-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.575605 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.618981 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.619022 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.620716 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5385fdd1-eb06-4d69-a383-236a55505108-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.620820 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5385fdd1-eb06-4d69-a383-236a55505108-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.620875 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d6wl\" (UniqueName: \"kubernetes.io/projected/5385fdd1-eb06-4d69-a383-236a55505108-kube-api-access-2d6wl\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.620892 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5385fdd1-eb06-4d69-a383-236a55505108-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.636868 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5385fdd1-eb06-4d69-a383-236a55505108-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.637026 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5385fdd1-eb06-4d69-a383-236a55505108-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.637592 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5385fdd1-eb06-4d69-a383-236a55505108-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.645334 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d6wl\" (UniqueName: \"kubernetes.io/projected/5385fdd1-eb06-4d69-a383-236a55505108-kube-api-access-2d6wl\") pod \"kube-state-metrics-0\" (UID: \"5385fdd1-eb06-4d69-a383-236a55505108\") " pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.649997 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b95fb5cb7-x5pbk"] Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.650247 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" podUID="9b6a2e16-8be3-4946-b625-545f052e86f0" containerName="dnsmasq-dns" containerID="cri-o://75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa" gracePeriod=10 Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.754051 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.947780 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.963696 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 08:36:40 crc kubenswrapper[4969]: I1004 08:36:40.963814 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.007231 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.077628 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e92558f-4f16-443e-b9a7-1c7f76013d2f" path="/var/lib/kubelet/pods/0e92558f-4f16-443e-b9a7-1c7f76013d2f/volumes" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.153617 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.232829 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-sb\") pod \"9b6a2e16-8be3-4946-b625-545f052e86f0\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.232882 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-svc\") pod \"9b6a2e16-8be3-4946-b625-545f052e86f0\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.232925 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvlq9\" (UniqueName: \"kubernetes.io/projected/9b6a2e16-8be3-4946-b625-545f052e86f0-kube-api-access-wvlq9\") pod \"9b6a2e16-8be3-4946-b625-545f052e86f0\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.233070 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-nb\") pod \"9b6a2e16-8be3-4946-b625-545f052e86f0\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.233170 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-config\") pod \"9b6a2e16-8be3-4946-b625-545f052e86f0\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.233219 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-swift-storage-0\") pod \"9b6a2e16-8be3-4946-b625-545f052e86f0\" (UID: \"9b6a2e16-8be3-4946-b625-545f052e86f0\") " Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.340732 4969 generic.go:334] "Generic (PLEG): container finished" podID="9b6a2e16-8be3-4946-b625-545f052e86f0" containerID="75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa" exitCode=0 Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.340800 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.340798 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" event={"ID":"9b6a2e16-8be3-4946-b625-545f052e86f0","Type":"ContainerDied","Data":"75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa"} Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.341166 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b95fb5cb7-x5pbk" event={"ID":"9b6a2e16-8be3-4946-b625-545f052e86f0","Type":"ContainerDied","Data":"98a8eebb705d7963f3fc6bb2ee15a275da93633596a6ceb8ef5da4693f9bd9b5"} Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.341193 4969 scope.go:117] "RemoveContainer" containerID="75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.344682 4969 generic.go:334] "Generic (PLEG): container finished" podID="cebc3b1f-5a32-4255-ae14-4fea09ad9f12" containerID="5c3ecbc26a9d98cfe05688d490ec149ed5852e06bd1bf98f08311d10f1aef34a" exitCode=0 Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.344774 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9hbl9" event={"ID":"cebc3b1f-5a32-4255-ae14-4fea09ad9f12","Type":"ContainerDied","Data":"5c3ecbc26a9d98cfe05688d490ec149ed5852e06bd1bf98f08311d10f1aef34a"} Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.395295 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.401982 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.402302 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="ceilometer-central-agent" containerID="cri-o://6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130" gracePeriod=30 Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.402474 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="proxy-httpd" containerID="cri-o://e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506" gracePeriod=30 Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.402531 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="sg-core" containerID="cri-o://930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee" gracePeriod=30 Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.402600 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="ceilometer-notification-agent" containerID="cri-o://9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec" gracePeriod=30 Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.411502 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-config" (OuterVolumeSpecName: "config") pod "9b6a2e16-8be3-4946-b625-545f052e86f0" (UID: "9b6a2e16-8be3-4946-b625-545f052e86f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.436665 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.445961 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 08:36:41 crc kubenswrapper[4969]: W1004 08:36:41.500353 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5385fdd1_eb06_4d69_a383_236a55505108.slice/crio-1d9f306eeacf947696682370cad40f8174237c423879cb48a2b44be2dfddfdd3 WatchSource:0}: Error finding container 1d9f306eeacf947696682370cad40f8174237c423879cb48a2b44be2dfddfdd3: Status 404 returned error can't find the container with id 1d9f306eeacf947696682370cad40f8174237c423879cb48a2b44be2dfddfdd3 Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.511463 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b6a2e16-8be3-4946-b625-545f052e86f0-kube-api-access-wvlq9" (OuterVolumeSpecName: "kube-api-access-wvlq9") pod "9b6a2e16-8be3-4946-b625-545f052e86f0" (UID: "9b6a2e16-8be3-4946-b625-545f052e86f0"). InnerVolumeSpecName "kube-api-access-wvlq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.538909 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvlq9\" (UniqueName: \"kubernetes.io/projected/9b6a2e16-8be3-4946-b625-545f052e86f0-kube-api-access-wvlq9\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.599535 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9b6a2e16-8be3-4946-b625-545f052e86f0" (UID: "9b6a2e16-8be3-4946-b625-545f052e86f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.600366 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9b6a2e16-8be3-4946-b625-545f052e86f0" (UID: "9b6a2e16-8be3-4946-b625-545f052e86f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.600855 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9b6a2e16-8be3-4946-b625-545f052e86f0" (UID: "9b6a2e16-8be3-4946-b625-545f052e86f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.601013 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9b6a2e16-8be3-4946-b625-545f052e86f0" (UID: "9b6a2e16-8be3-4946-b625-545f052e86f0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.626382 4969 scope.go:117] "RemoveContainer" containerID="683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.641229 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.641827 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.642213 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.642279 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b6a2e16-8be3-4946-b625-545f052e86f0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.707487 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.707879 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.741834 4969 scope.go:117] "RemoveContainer" containerID="75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa" Oct 04 08:36:41 crc kubenswrapper[4969]: E1004 08:36:41.742329 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa\": container with ID starting with 75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa not found: ID does not exist" containerID="75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.742404 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa"} err="failed to get container status \"75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa\": rpc error: code = NotFound desc = could not find container \"75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa\": container with ID starting with 75867843796d1913214e00fc82030cda09f3f438b310334ca2b36f5ba4e9fbaa not found: ID does not exist" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.742499 4969 scope.go:117] "RemoveContainer" containerID="683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502" Oct 04 08:36:41 crc kubenswrapper[4969]: E1004 08:36:41.742780 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502\": container with ID starting with 683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502 not found: ID does not exist" containerID="683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.742822 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502"} err="failed to get container status \"683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502\": rpc error: code = NotFound desc = could not find container \"683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502\": container with ID starting with 683d2adb32dbe63e605773c0443744c28130bb3098fac69702edd48058de0502 not found: ID does not exist" Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.788874 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b95fb5cb7-x5pbk"] Oct 04 08:36:41 crc kubenswrapper[4969]: I1004 08:36:41.798816 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b95fb5cb7-x5pbk"] Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.355590 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5385fdd1-eb06-4d69-a383-236a55505108","Type":"ContainerStarted","Data":"5e054d354e66ba00862f86bb82d3f29d788ab3ddd8b961d059a9408c08ef2667"} Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.355639 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5385fdd1-eb06-4d69-a383-236a55505108","Type":"ContainerStarted","Data":"1d9f306eeacf947696682370cad40f8174237c423879cb48a2b44be2dfddfdd3"} Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.355783 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.363357 4969 generic.go:334] "Generic (PLEG): container finished" podID="f0021ab7-5417-4f92-a69b-eed20068b066" containerID="e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506" exitCode=0 Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.363397 4969 generic.go:334] "Generic (PLEG): container finished" podID="f0021ab7-5417-4f92-a69b-eed20068b066" containerID="930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee" exitCode=2 Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.363405 4969 generic.go:334] "Generic (PLEG): container finished" podID="f0021ab7-5417-4f92-a69b-eed20068b066" containerID="6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130" exitCode=0 Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.364228 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerDied","Data":"e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506"} Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.364257 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerDied","Data":"930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee"} Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.364268 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerDied","Data":"6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130"} Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.382511 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.9642449530000001 podStartE2EDuration="2.382487652s" podCreationTimestamp="2025-10-04 08:36:40 +0000 UTC" firstStartedPulling="2025-10-04 08:36:41.503956235 +0000 UTC m=+1229.258225049" lastFinishedPulling="2025-10-04 08:36:41.922198944 +0000 UTC m=+1229.676467748" observedRunningTime="2025-10-04 08:36:42.37911178 +0000 UTC m=+1230.133380604" watchObservedRunningTime="2025-10-04 08:36:42.382487652 +0000 UTC m=+1230.136756466" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.718954 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.765800 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-scripts\") pod \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.765899 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-combined-ca-bundle\") pod \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.766060 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j72nj\" (UniqueName: \"kubernetes.io/projected/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-kube-api-access-j72nj\") pod \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.766107 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-config-data\") pod \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\" (UID: \"cebc3b1f-5a32-4255-ae14-4fea09ad9f12\") " Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.770518 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-kube-api-access-j72nj" (OuterVolumeSpecName: "kube-api-access-j72nj") pod "cebc3b1f-5a32-4255-ae14-4fea09ad9f12" (UID: "cebc3b1f-5a32-4255-ae14-4fea09ad9f12"). InnerVolumeSpecName "kube-api-access-j72nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.772257 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-scripts" (OuterVolumeSpecName: "scripts") pod "cebc3b1f-5a32-4255-ae14-4fea09ad9f12" (UID: "cebc3b1f-5a32-4255-ae14-4fea09ad9f12"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.796683 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-config-data" (OuterVolumeSpecName: "config-data") pod "cebc3b1f-5a32-4255-ae14-4fea09ad9f12" (UID: "cebc3b1f-5a32-4255-ae14-4fea09ad9f12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.816465 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cebc3b1f-5a32-4255-ae14-4fea09ad9f12" (UID: "cebc3b1f-5a32-4255-ae14-4fea09ad9f12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.869591 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j72nj\" (UniqueName: \"kubernetes.io/projected/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-kube-api-access-j72nj\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.869623 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.869633 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:42 crc kubenswrapper[4969]: I1004 08:36:42.869641 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebc3b1f-5a32-4255-ae14-4fea09ad9f12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.065924 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b6a2e16-8be3-4946-b625-545f052e86f0" path="/var/lib/kubelet/pods/9b6a2e16-8be3-4946-b625-545f052e86f0/volumes" Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.378105 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9hbl9" Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.379187 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9hbl9" event={"ID":"cebc3b1f-5a32-4255-ae14-4fea09ad9f12","Type":"ContainerDied","Data":"5fadb0a2b6a46c181240deea19ccf779527b5d67d1559c0407192eed8c75e94c"} Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.379229 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fadb0a2b6a46c181240deea19ccf779527b5d67d1559c0407192eed8c75e94c" Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.561659 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.561916 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-log" containerID="cri-o://f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff" gracePeriod=30 Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.562100 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-api" containerID="cri-o://b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730" gracePeriod=30 Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.578955 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.606853 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.607572 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerName="nova-metadata-log" containerID="cri-o://533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811" gracePeriod=30 Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.609002 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerName="nova-metadata-metadata" containerID="cri-o://074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570" gracePeriod=30 Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.682772 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.716044 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.720380 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 08:36:43 crc kubenswrapper[4969]: I1004 08:36:43.720436 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.192079 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.301596 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-config-data\") pod \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.301677 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjmn5\" (UniqueName: \"kubernetes.io/projected/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-kube-api-access-mjmn5\") pod \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.301723 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-nova-metadata-tls-certs\") pod \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.301751 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-logs\") pod \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.301806 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-combined-ca-bundle\") pod \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\" (UID: \"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2\") " Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.302110 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-logs" (OuterVolumeSpecName: "logs") pod "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" (UID: "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.302271 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.306558 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-kube-api-access-mjmn5" (OuterVolumeSpecName: "kube-api-access-mjmn5") pod "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" (UID: "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2"). InnerVolumeSpecName "kube-api-access-mjmn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.337231 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" (UID: "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.340777 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-config-data" (OuterVolumeSpecName: "config-data") pod "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" (UID: "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.359763 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" (UID: "91ad8cc9-ec8b-4a24-a98f-036ec933a7b2"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.388308 4969 generic.go:334] "Generic (PLEG): container finished" podID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerID="074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570" exitCode=0 Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.388341 4969 generic.go:334] "Generic (PLEG): container finished" podID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerID="533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811" exitCode=143 Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.388384 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2","Type":"ContainerDied","Data":"074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570"} Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.388411 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2","Type":"ContainerDied","Data":"533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811"} Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.388436 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91ad8cc9-ec8b-4a24-a98f-036ec933a7b2","Type":"ContainerDied","Data":"a1e91b370b353ed4091f886237a6f9bb30927dbf531a6e98cd60c975ca5d0630"} Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.388452 4969 scope.go:117] "RemoveContainer" containerID="074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.388616 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.395660 4969 generic.go:334] "Generic (PLEG): container finished" podID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerID="f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff" exitCode=143 Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.395844 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" containerName="nova-scheduler-scheduler" containerID="cri-o://9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7" gracePeriod=30 Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.396175 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589f782e-3da1-4dd5-b652-672955ea4bbc","Type":"ContainerDied","Data":"f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff"} Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.396878 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.404333 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.404361 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.404372 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjmn5\" (UniqueName: \"kubernetes.io/projected/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-kube-api-access-mjmn5\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.404736 4969 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.421691 4969 scope.go:117] "RemoveContainer" containerID="533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.440567 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.448867 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.454187 4969 scope.go:117] "RemoveContainer" containerID="074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570" Oct 04 08:36:44 crc kubenswrapper[4969]: E1004 08:36:44.454831 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570\": container with ID starting with 074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570 not found: ID does not exist" containerID="074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.454884 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570"} err="failed to get container status \"074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570\": rpc error: code = NotFound desc = could not find container \"074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570\": container with ID starting with 074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570 not found: ID does not exist" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.454918 4969 scope.go:117] "RemoveContainer" containerID="533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811" Oct 04 08:36:44 crc kubenswrapper[4969]: E1004 08:36:44.455224 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811\": container with ID starting with 533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811 not found: ID does not exist" containerID="533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.455487 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811"} err="failed to get container status \"533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811\": rpc error: code = NotFound desc = could not find container \"533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811\": container with ID starting with 533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811 not found: ID does not exist" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.455649 4969 scope.go:117] "RemoveContainer" containerID="074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.459706 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.460946 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570"} err="failed to get container status \"074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570\": rpc error: code = NotFound desc = could not find container \"074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570\": container with ID starting with 074385cad71c0160f033c97054528fa4e40b9025720171dbbb10b4080546c570 not found: ID does not exist" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.461055 4969 scope.go:117] "RemoveContainer" containerID="533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.462389 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811"} err="failed to get container status \"533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811\": rpc error: code = NotFound desc = could not find container \"533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811\": container with ID starting with 533f47a3bc41ff9ad973b360509487db896aadc529a5f61eaf9451188db89811 not found: ID does not exist" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.496473 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:44 crc kubenswrapper[4969]: E1004 08:36:44.497105 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerName="nova-metadata-metadata" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.497203 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerName="nova-metadata-metadata" Oct 04 08:36:44 crc kubenswrapper[4969]: E1004 08:36:44.497264 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b6a2e16-8be3-4946-b625-545f052e86f0" containerName="dnsmasq-dns" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.497320 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b6a2e16-8be3-4946-b625-545f052e86f0" containerName="dnsmasq-dns" Oct 04 08:36:44 crc kubenswrapper[4969]: E1004 08:36:44.497382 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cebc3b1f-5a32-4255-ae14-4fea09ad9f12" containerName="nova-manage" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.497457 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="cebc3b1f-5a32-4255-ae14-4fea09ad9f12" containerName="nova-manage" Oct 04 08:36:44 crc kubenswrapper[4969]: E1004 08:36:44.497515 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b6a2e16-8be3-4946-b625-545f052e86f0" containerName="init" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.497574 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b6a2e16-8be3-4946-b625-545f052e86f0" containerName="init" Oct 04 08:36:44 crc kubenswrapper[4969]: E1004 08:36:44.497650 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerName="nova-metadata-log" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.497699 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerName="nova-metadata-log" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.497928 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerName="nova-metadata-log" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.498001 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b6a2e16-8be3-4946-b625-545f052e86f0" containerName="dnsmasq-dns" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.498061 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="cebc3b1f-5a32-4255-ae14-4fea09ad9f12" containerName="nova-manage" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.498131 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" containerName="nova-metadata-metadata" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.499203 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.511717 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.511934 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.512010 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.607460 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-config-data\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.607715 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.607813 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.608123 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-logs\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.608263 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x722f\" (UniqueName: \"kubernetes.io/projected/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-kube-api-access-x722f\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.710631 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.710795 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.710988 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-logs\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.711127 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x722f\" (UniqueName: \"kubernetes.io/projected/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-kube-api-access-x722f\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.711310 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-config-data\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.711523 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-logs\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.713751 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.714886 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.715711 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-config-data\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.726993 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x722f\" (UniqueName: \"kubernetes.io/projected/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-kube-api-access-x722f\") pod \"nova-metadata-0\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " pod="openstack/nova-metadata-0" Oct 04 08:36:44 crc kubenswrapper[4969]: I1004 08:36:44.822450 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:36:45 crc kubenswrapper[4969]: I1004 08:36:45.070279 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91ad8cc9-ec8b-4a24-a98f-036ec933a7b2" path="/var/lib/kubelet/pods/91ad8cc9-ec8b-4a24-a98f-036ec933a7b2/volumes" Oct 04 08:36:45 crc kubenswrapper[4969]: I1004 08:36:45.369727 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:36:45 crc kubenswrapper[4969]: I1004 08:36:45.424996 4969 generic.go:334] "Generic (PLEG): container finished" podID="2160e9f1-8d10-4ddf-931f-972f80fed6ba" containerID="906c11d6be4da220c5de9e0d3c69f0a215ef98e2509527008b4d4ab455a32467" exitCode=0 Oct 04 08:36:45 crc kubenswrapper[4969]: I1004 08:36:45.425064 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9xhfr" event={"ID":"2160e9f1-8d10-4ddf-931f-972f80fed6ba","Type":"ContainerDied","Data":"906c11d6be4da220c5de9e0d3c69f0a215ef98e2509527008b4d4ab455a32467"} Oct 04 08:36:45 crc kubenswrapper[4969]: I1004 08:36:45.429954 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"788a64ad-f6c3-46e5-941f-4b5f1fe568c7","Type":"ContainerStarted","Data":"668f0417bff4817822ffe4c4ff8f063d04728ea8bc0b09ebd904cabad28b6e45"} Oct 04 08:36:45 crc kubenswrapper[4969]: E1004 08:36:45.966450 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 08:36:45 crc kubenswrapper[4969]: E1004 08:36:45.968205 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 08:36:45 crc kubenswrapper[4969]: E1004 08:36:45.970102 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 08:36:45 crc kubenswrapper[4969]: E1004 08:36:45.970128 4969 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" containerName="nova-scheduler-scheduler" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.233443 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.342476 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-combined-ca-bundle\") pod \"589f782e-3da1-4dd5-b652-672955ea4bbc\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.342574 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589f782e-3da1-4dd5-b652-672955ea4bbc-logs\") pod \"589f782e-3da1-4dd5-b652-672955ea4bbc\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.342714 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txwz2\" (UniqueName: \"kubernetes.io/projected/589f782e-3da1-4dd5-b652-672955ea4bbc-kube-api-access-txwz2\") pod \"589f782e-3da1-4dd5-b652-672955ea4bbc\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.342803 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-config-data\") pod \"589f782e-3da1-4dd5-b652-672955ea4bbc\" (UID: \"589f782e-3da1-4dd5-b652-672955ea4bbc\") " Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.344200 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/589f782e-3da1-4dd5-b652-672955ea4bbc-logs" (OuterVolumeSpecName: "logs") pod "589f782e-3da1-4dd5-b652-672955ea4bbc" (UID: "589f782e-3da1-4dd5-b652-672955ea4bbc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.348092 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/589f782e-3da1-4dd5-b652-672955ea4bbc-kube-api-access-txwz2" (OuterVolumeSpecName: "kube-api-access-txwz2") pod "589f782e-3da1-4dd5-b652-672955ea4bbc" (UID: "589f782e-3da1-4dd5-b652-672955ea4bbc"). InnerVolumeSpecName "kube-api-access-txwz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.369761 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-config-data" (OuterVolumeSpecName: "config-data") pod "589f782e-3da1-4dd5-b652-672955ea4bbc" (UID: "589f782e-3da1-4dd5-b652-672955ea4bbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.398292 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "589f782e-3da1-4dd5-b652-672955ea4bbc" (UID: "589f782e-3da1-4dd5-b652-672955ea4bbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.440866 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"788a64ad-f6c3-46e5-941f-4b5f1fe568c7","Type":"ContainerStarted","Data":"64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7"} Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.440918 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"788a64ad-f6c3-46e5-941f-4b5f1fe568c7","Type":"ContainerStarted","Data":"99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500"} Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.443088 4969 generic.go:334] "Generic (PLEG): container finished" podID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerID="b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730" exitCode=0 Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.443254 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.443351 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589f782e-3da1-4dd5-b652-672955ea4bbc","Type":"ContainerDied","Data":"b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730"} Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.443369 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589f782e-3da1-4dd5-b652-672955ea4bbc","Type":"ContainerDied","Data":"e049c26d48397831bd3ecef749e541966b8ef6ea1efb4df9a3aca8c674e2c6f8"} Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.443385 4969 scope.go:117] "RemoveContainer" containerID="b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.445168 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589f782e-3da1-4dd5-b652-672955ea4bbc-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.445189 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txwz2\" (UniqueName: \"kubernetes.io/projected/589f782e-3da1-4dd5-b652-672955ea4bbc-kube-api-access-txwz2\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.445200 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.445209 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589f782e-3da1-4dd5-b652-672955ea4bbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.483007 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.482961777 podStartE2EDuration="2.482961777s" podCreationTimestamp="2025-10-04 08:36:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:46.46536448 +0000 UTC m=+1234.219633294" watchObservedRunningTime="2025-10-04 08:36:46.482961777 +0000 UTC m=+1234.237230591" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.484635 4969 scope.go:117] "RemoveContainer" containerID="f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.505046 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.516447 4969 scope.go:117] "RemoveContainer" containerID="b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730" Oct 04 08:36:46 crc kubenswrapper[4969]: E1004 08:36:46.517009 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730\": container with ID starting with b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730 not found: ID does not exist" containerID="b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.517050 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730"} err="failed to get container status \"b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730\": rpc error: code = NotFound desc = could not find container \"b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730\": container with ID starting with b798922aefc4dd5a7061672458f17506a1852b40302cc0030399f54a090e1730 not found: ID does not exist" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.517078 4969 scope.go:117] "RemoveContainer" containerID="f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff" Oct 04 08:36:46 crc kubenswrapper[4969]: E1004 08:36:46.517837 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff\": container with ID starting with f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff not found: ID does not exist" containerID="f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.517904 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff"} err="failed to get container status \"f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff\": rpc error: code = NotFound desc = could not find container \"f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff\": container with ID starting with f8a66ca94254fa1e87dc264960601ff967397ac65a870aae2ea19c1ac30fe1ff not found: ID does not exist" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.548791 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.563519 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 08:36:46 crc kubenswrapper[4969]: E1004 08:36:46.564077 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-log" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.564100 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-log" Oct 04 08:36:46 crc kubenswrapper[4969]: E1004 08:36:46.564123 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-api" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.564132 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-api" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.564403 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-log" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.564485 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" containerName="nova-api-api" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.577971 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.580956 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.596168 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.649968 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.650018 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/451dbcc5-7ad5-429a-8404-73b2c75466bd-logs\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.650156 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-config-data\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.650188 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vzgb\" (UniqueName: \"kubernetes.io/projected/451dbcc5-7ad5-429a-8404-73b2c75466bd-kube-api-access-2vzgb\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.751740 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-config-data\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.752389 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vzgb\" (UniqueName: \"kubernetes.io/projected/451dbcc5-7ad5-429a-8404-73b2c75466bd-kube-api-access-2vzgb\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.753208 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.753257 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/451dbcc5-7ad5-429a-8404-73b2c75466bd-logs\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.753730 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/451dbcc5-7ad5-429a-8404-73b2c75466bd-logs\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.758679 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-config-data\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.759019 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.780600 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vzgb\" (UniqueName: \"kubernetes.io/projected/451dbcc5-7ad5-429a-8404-73b2c75466bd-kube-api-access-2vzgb\") pod \"nova-api-0\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.893569 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:36:46 crc kubenswrapper[4969]: I1004 08:36:46.983904 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.058749 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-combined-ca-bundle\") pod \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.059017 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb92c\" (UniqueName: \"kubernetes.io/projected/2160e9f1-8d10-4ddf-931f-972f80fed6ba-kube-api-access-vb92c\") pod \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.059067 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-scripts\") pod \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.059181 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-config-data\") pod \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\" (UID: \"2160e9f1-8d10-4ddf-931f-972f80fed6ba\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.063880 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2160e9f1-8d10-4ddf-931f-972f80fed6ba-kube-api-access-vb92c" (OuterVolumeSpecName: "kube-api-access-vb92c") pod "2160e9f1-8d10-4ddf-931f-972f80fed6ba" (UID: "2160e9f1-8d10-4ddf-931f-972f80fed6ba"). InnerVolumeSpecName "kube-api-access-vb92c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.066676 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-scripts" (OuterVolumeSpecName: "scripts") pod "2160e9f1-8d10-4ddf-931f-972f80fed6ba" (UID: "2160e9f1-8d10-4ddf-931f-972f80fed6ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.071161 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="589f782e-3da1-4dd5-b652-672955ea4bbc" path="/var/lib/kubelet/pods/589f782e-3da1-4dd5-b652-672955ea4bbc/volumes" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.086945 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-config-data" (OuterVolumeSpecName: "config-data") pod "2160e9f1-8d10-4ddf-931f-972f80fed6ba" (UID: "2160e9f1-8d10-4ddf-931f-972f80fed6ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.094552 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2160e9f1-8d10-4ddf-931f-972f80fed6ba" (UID: "2160e9f1-8d10-4ddf-931f-972f80fed6ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.130880 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.162037 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.162090 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb92c\" (UniqueName: \"kubernetes.io/projected/2160e9f1-8d10-4ddf-931f-972f80fed6ba-kube-api-access-vb92c\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.162106 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.162119 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2160e9f1-8d10-4ddf-931f-972f80fed6ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.263325 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-scripts\") pod \"f0021ab7-5417-4f92-a69b-eed20068b066\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.263373 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-log-httpd\") pod \"f0021ab7-5417-4f92-a69b-eed20068b066\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.263395 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-run-httpd\") pod \"f0021ab7-5417-4f92-a69b-eed20068b066\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.263498 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-config-data\") pod \"f0021ab7-5417-4f92-a69b-eed20068b066\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.263553 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lz4w\" (UniqueName: \"kubernetes.io/projected/f0021ab7-5417-4f92-a69b-eed20068b066-kube-api-access-8lz4w\") pod \"f0021ab7-5417-4f92-a69b-eed20068b066\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.263627 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-sg-core-conf-yaml\") pod \"f0021ab7-5417-4f92-a69b-eed20068b066\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.263711 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-combined-ca-bundle\") pod \"f0021ab7-5417-4f92-a69b-eed20068b066\" (UID: \"f0021ab7-5417-4f92-a69b-eed20068b066\") " Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.264024 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f0021ab7-5417-4f92-a69b-eed20068b066" (UID: "f0021ab7-5417-4f92-a69b-eed20068b066"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.264071 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f0021ab7-5417-4f92-a69b-eed20068b066" (UID: "f0021ab7-5417-4f92-a69b-eed20068b066"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.264600 4969 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.264623 4969 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0021ab7-5417-4f92-a69b-eed20068b066-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.267943 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-scripts" (OuterVolumeSpecName: "scripts") pod "f0021ab7-5417-4f92-a69b-eed20068b066" (UID: "f0021ab7-5417-4f92-a69b-eed20068b066"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.269273 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0021ab7-5417-4f92-a69b-eed20068b066-kube-api-access-8lz4w" (OuterVolumeSpecName: "kube-api-access-8lz4w") pod "f0021ab7-5417-4f92-a69b-eed20068b066" (UID: "f0021ab7-5417-4f92-a69b-eed20068b066"). InnerVolumeSpecName "kube-api-access-8lz4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.292026 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f0021ab7-5417-4f92-a69b-eed20068b066" (UID: "f0021ab7-5417-4f92-a69b-eed20068b066"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.360585 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0021ab7-5417-4f92-a69b-eed20068b066" (UID: "f0021ab7-5417-4f92-a69b-eed20068b066"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.367810 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.368268 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lz4w\" (UniqueName: \"kubernetes.io/projected/f0021ab7-5417-4f92-a69b-eed20068b066-kube-api-access-8lz4w\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.368294 4969 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.368312 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.385301 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-config-data" (OuterVolumeSpecName: "config-data") pod "f0021ab7-5417-4f92-a69b-eed20068b066" (UID: "f0021ab7-5417-4f92-a69b-eed20068b066"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:47 crc kubenswrapper[4969]: W1004 08:36:47.387625 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod451dbcc5_7ad5_429a_8404_73b2c75466bd.slice/crio-485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f WatchSource:0}: Error finding container 485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f: Status 404 returned error can't find the container with id 485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.393625 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.465554 4969 generic.go:334] "Generic (PLEG): container finished" podID="f0021ab7-5417-4f92-a69b-eed20068b066" containerID="9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec" exitCode=0 Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.465654 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.465635 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerDied","Data":"9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec"} Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.466857 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0021ab7-5417-4f92-a69b-eed20068b066","Type":"ContainerDied","Data":"7704c1f08b9307cc7a7945473fcd943c1ae74113aa4e0990ccb9b3519809481d"} Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.466909 4969 scope.go:117] "RemoveContainer" containerID="e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.470262 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0021ab7-5417-4f92-a69b-eed20068b066-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.470741 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9xhfr" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.471249 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9xhfr" event={"ID":"2160e9f1-8d10-4ddf-931f-972f80fed6ba","Type":"ContainerDied","Data":"92e5c94264bec744e06c0bd7cfa1ccfd83d213a99816a217c7ee5e9a1bc2300c"} Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.471325 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92e5c94264bec744e06c0bd7cfa1ccfd83d213a99816a217c7ee5e9a1bc2300c" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.479490 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"451dbcc5-7ad5-429a-8404-73b2c75466bd","Type":"ContainerStarted","Data":"485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f"} Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.511327 4969 scope.go:117] "RemoveContainer" containerID="930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.522777 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.538589 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.546774 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 08:36:47 crc kubenswrapper[4969]: E1004 08:36:47.547306 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="sg-core" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547329 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="sg-core" Oct 04 08:36:47 crc kubenswrapper[4969]: E1004 08:36:47.547359 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="proxy-httpd" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547368 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="proxy-httpd" Oct 04 08:36:47 crc kubenswrapper[4969]: E1004 08:36:47.547396 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="ceilometer-notification-agent" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547405 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="ceilometer-notification-agent" Oct 04 08:36:47 crc kubenswrapper[4969]: E1004 08:36:47.547438 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2160e9f1-8d10-4ddf-931f-972f80fed6ba" containerName="nova-cell1-conductor-db-sync" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547450 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="2160e9f1-8d10-4ddf-931f-972f80fed6ba" containerName="nova-cell1-conductor-db-sync" Oct 04 08:36:47 crc kubenswrapper[4969]: E1004 08:36:47.547478 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="ceilometer-central-agent" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547486 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="ceilometer-central-agent" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547703 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="proxy-httpd" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547741 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="sg-core" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547756 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="2160e9f1-8d10-4ddf-931f-972f80fed6ba" containerName="nova-cell1-conductor-db-sync" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547769 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="ceilometer-central-agent" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.547785 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" containerName="ceilometer-notification-agent" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.548009 4969 scope.go:117] "RemoveContainer" containerID="9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.548634 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.556761 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.568002 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.572852 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.572962 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.576438 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.576722 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.576887 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.578053 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.582759 4969 scope.go:117] "RemoveContainer" containerID="6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.623368 4969 scope.go:117] "RemoveContainer" containerID="e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506" Oct 04 08:36:47 crc kubenswrapper[4969]: E1004 08:36:47.624670 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506\": container with ID starting with e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506 not found: ID does not exist" containerID="e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.624716 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506"} err="failed to get container status \"e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506\": rpc error: code = NotFound desc = could not find container \"e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506\": container with ID starting with e2e2435c5e928b61f65d85c4d4c1c6ccdd37a933b39460e6fa43e9dff179a506 not found: ID does not exist" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.624741 4969 scope.go:117] "RemoveContainer" containerID="930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee" Oct 04 08:36:47 crc kubenswrapper[4969]: E1004 08:36:47.625163 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee\": container with ID starting with 930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee not found: ID does not exist" containerID="930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.625187 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee"} err="failed to get container status \"930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee\": rpc error: code = NotFound desc = could not find container \"930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee\": container with ID starting with 930e3e84f77e1405e89c490a8bf21381a1f57afbf7921f89d81c914e5499ffee not found: ID does not exist" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.625201 4969 scope.go:117] "RemoveContainer" containerID="9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec" Oct 04 08:36:47 crc kubenswrapper[4969]: E1004 08:36:47.625485 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec\": container with ID starting with 9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec not found: ID does not exist" containerID="9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.625505 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec"} err="failed to get container status \"9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec\": rpc error: code = NotFound desc = could not find container \"9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec\": container with ID starting with 9a4f47a4b83c587367bc0b489d2b0de4bcfb3de10ed5f3417096dd0f9e2184ec not found: ID does not exist" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.625517 4969 scope.go:117] "RemoveContainer" containerID="6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130" Oct 04 08:36:47 crc kubenswrapper[4969]: E1004 08:36:47.626450 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130\": container with ID starting with 6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130 not found: ID does not exist" containerID="6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.626479 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130"} err="failed to get container status \"6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130\": rpc error: code = NotFound desc = could not find container \"6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130\": container with ID starting with 6e7b85c2babd8c788948e8efa1a2963cad3a333fb6825530f114c12163417130 not found: ID does not exist" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.675230 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-scripts\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.675320 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac48f46b-0304-424d-b394-a1e2a274230a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ac48f46b-0304-424d-b394-a1e2a274230a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.675511 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac48f46b-0304-424d-b394-a1e2a274230a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ac48f46b-0304-424d-b394-a1e2a274230a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.675545 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-run-httpd\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.675574 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-config-data\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.675602 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvkg2\" (UniqueName: \"kubernetes.io/projected/ac48f46b-0304-424d-b394-a1e2a274230a-kube-api-access-pvkg2\") pod \"nova-cell1-conductor-0\" (UID: \"ac48f46b-0304-424d-b394-a1e2a274230a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.675739 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-log-httpd\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.676044 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.676080 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.676095 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.676113 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7xsm\" (UniqueName: \"kubernetes.io/projected/bb36ec4c-072a-40b9-bd4d-a23938c71a67-kube-api-access-z7xsm\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778395 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvkg2\" (UniqueName: \"kubernetes.io/projected/ac48f46b-0304-424d-b394-a1e2a274230a-kube-api-access-pvkg2\") pod \"nova-cell1-conductor-0\" (UID: \"ac48f46b-0304-424d-b394-a1e2a274230a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778469 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-log-httpd\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778509 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778526 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778542 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778558 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7xsm\" (UniqueName: \"kubernetes.io/projected/bb36ec4c-072a-40b9-bd4d-a23938c71a67-kube-api-access-z7xsm\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778612 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-scripts\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778650 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac48f46b-0304-424d-b394-a1e2a274230a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ac48f46b-0304-424d-b394-a1e2a274230a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778698 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac48f46b-0304-424d-b394-a1e2a274230a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ac48f46b-0304-424d-b394-a1e2a274230a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778715 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-run-httpd\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.778732 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-config-data\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.781299 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-run-httpd\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.783070 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.783836 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-config-data\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.783996 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac48f46b-0304-424d-b394-a1e2a274230a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ac48f46b-0304-424d-b394-a1e2a274230a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.784135 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac48f46b-0304-424d-b394-a1e2a274230a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ac48f46b-0304-424d-b394-a1e2a274230a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.785592 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.785662 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-scripts\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.786112 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.788857 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-log-httpd\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.796576 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7xsm\" (UniqueName: \"kubernetes.io/projected/bb36ec4c-072a-40b9-bd4d-a23938c71a67-kube-api-access-z7xsm\") pod \"ceilometer-0\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " pod="openstack/ceilometer-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.796644 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvkg2\" (UniqueName: \"kubernetes.io/projected/ac48f46b-0304-424d-b394-a1e2a274230a-kube-api-access-pvkg2\") pod \"nova-cell1-conductor-0\" (UID: \"ac48f46b-0304-424d-b394-a1e2a274230a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.893855 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:47 crc kubenswrapper[4969]: I1004 08:36:47.913237 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:36:48 crc kubenswrapper[4969]: I1004 08:36:48.343095 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 08:36:48 crc kubenswrapper[4969]: W1004 08:36:48.350454 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac48f46b_0304_424d_b394_a1e2a274230a.slice/crio-5ad30bef7532f69844df9794fc57f6e775da50d210c95c0cd9cea9d3f2752480 WatchSource:0}: Error finding container 5ad30bef7532f69844df9794fc57f6e775da50d210c95c0cd9cea9d3f2752480: Status 404 returned error can't find the container with id 5ad30bef7532f69844df9794fc57f6e775da50d210c95c0cd9cea9d3f2752480 Oct 04 08:36:48 crc kubenswrapper[4969]: I1004 08:36:48.475617 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:36:48 crc kubenswrapper[4969]: I1004 08:36:48.495459 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ac48f46b-0304-424d-b394-a1e2a274230a","Type":"ContainerStarted","Data":"5ad30bef7532f69844df9794fc57f6e775da50d210c95c0cd9cea9d3f2752480"} Oct 04 08:36:48 crc kubenswrapper[4969]: W1004 08:36:48.504372 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb36ec4c_072a_40b9_bd4d_a23938c71a67.slice/crio-f2fe71499549387275bf3dc8d66d2c646e699d14ba088e11c401c16ffeed89df WatchSource:0}: Error finding container f2fe71499549387275bf3dc8d66d2c646e699d14ba088e11c401c16ffeed89df: Status 404 returned error can't find the container with id f2fe71499549387275bf3dc8d66d2c646e699d14ba088e11c401c16ffeed89df Oct 04 08:36:48 crc kubenswrapper[4969]: I1004 08:36:48.505435 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"451dbcc5-7ad5-429a-8404-73b2c75466bd","Type":"ContainerStarted","Data":"619b585ea6b3e97697dda5987967b64d34cc5bb22638bf518754b40debef639f"} Oct 04 08:36:48 crc kubenswrapper[4969]: I1004 08:36:48.505480 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"451dbcc5-7ad5-429a-8404-73b2c75466bd","Type":"ContainerStarted","Data":"a55d3802ea7dc76dcd28478a0d3a7f6944d37e3892d2b4ae2983a5a0ad86c261"} Oct 04 08:36:48 crc kubenswrapper[4969]: I1004 08:36:48.529590 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.5295706559999998 podStartE2EDuration="2.529570656s" podCreationTimestamp="2025-10-04 08:36:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:48.52815424 +0000 UTC m=+1236.282423054" watchObservedRunningTime="2025-10-04 08:36:48.529570656 +0000 UTC m=+1236.283839490" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.062673 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.066617 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0021ab7-5417-4f92-a69b-eed20068b066" path="/var/lib/kubelet/pods/f0021ab7-5417-4f92-a69b-eed20068b066/volumes" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.255068 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-combined-ca-bundle\") pod \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.255415 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjrdk\" (UniqueName: \"kubernetes.io/projected/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-kube-api-access-zjrdk\") pod \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.255576 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-config-data\") pod \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\" (UID: \"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe\") " Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.260366 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-kube-api-access-zjrdk" (OuterVolumeSpecName: "kube-api-access-zjrdk") pod "25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" (UID: "25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe"). InnerVolumeSpecName "kube-api-access-zjrdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.289555 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" (UID: "25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.304297 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-config-data" (OuterVolumeSpecName: "config-data") pod "25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" (UID: "25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.358505 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.358562 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.358584 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjrdk\" (UniqueName: \"kubernetes.io/projected/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe-kube-api-access-zjrdk\") on node \"crc\" DevicePath \"\"" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.518779 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ac48f46b-0304-424d-b394-a1e2a274230a","Type":"ContainerStarted","Data":"c8dd94be6f30620e2faf9f0c5bd4d2747b77389a20aa28071535e0dc222f905d"} Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.518965 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.520543 4969 generic.go:334] "Generic (PLEG): container finished" podID="25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" containerID="9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7" exitCode=0 Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.520603 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe","Type":"ContainerDied","Data":"9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7"} Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.520622 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe","Type":"ContainerDied","Data":"da0a739eb7373234ae079dfe2a2a6ddc3b4cb6905b5a5ce78684652a0b0c5779"} Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.520622 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.520639 4969 scope.go:117] "RemoveContainer" containerID="9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.524180 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerStarted","Data":"3436b85a700dab71576f2b84084f24b062c0d248728d5ca6871242ea7986d689"} Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.524209 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerStarted","Data":"8a7f7c9681d88161014fb95f54a27fed0435322438935efcda4191c686af1022"} Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.524219 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerStarted","Data":"f2fe71499549387275bf3dc8d66d2c646e699d14ba088e11c401c16ffeed89df"} Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.568509 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.56848852 podStartE2EDuration="2.56848852s" podCreationTimestamp="2025-10-04 08:36:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:49.554732456 +0000 UTC m=+1237.309001280" watchObservedRunningTime="2025-10-04 08:36:49.56848852 +0000 UTC m=+1237.322757344" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.568685 4969 scope.go:117] "RemoveContainer" containerID="9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7" Oct 04 08:36:49 crc kubenswrapper[4969]: E1004 08:36:49.577588 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7\": container with ID starting with 9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7 not found: ID does not exist" containerID="9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.577619 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7"} err="failed to get container status \"9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7\": rpc error: code = NotFound desc = could not find container \"9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7\": container with ID starting with 9448f5e608550941998aa2de1c18e951e93c4aac056cd239dcec189aadfaacc7 not found: ID does not exist" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.585693 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.600258 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.623732 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:36:49 crc kubenswrapper[4969]: E1004 08:36:49.624267 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" containerName="nova-scheduler-scheduler" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.624290 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" containerName="nova-scheduler-scheduler" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.624639 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" containerName="nova-scheduler-scheduler" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.625510 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.627477 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.637196 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.767664 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-config-data\") pod \"nova-scheduler-0\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.768002 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.768094 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sthj\" (UniqueName: \"kubernetes.io/projected/434f8bca-e752-49f3-9bce-286c90b64c00-kube-api-access-4sthj\") pod \"nova-scheduler-0\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.822811 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.822853 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.870628 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-config-data\") pod \"nova-scheduler-0\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.870679 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.870758 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sthj\" (UniqueName: \"kubernetes.io/projected/434f8bca-e752-49f3-9bce-286c90b64c00-kube-api-access-4sthj\") pod \"nova-scheduler-0\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.876226 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-config-data\") pod \"nova-scheduler-0\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.883108 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.903213 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sthj\" (UniqueName: \"kubernetes.io/projected/434f8bca-e752-49f3-9bce-286c90b64c00-kube-api-access-4sthj\") pod \"nova-scheduler-0\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " pod="openstack/nova-scheduler-0" Oct 04 08:36:49 crc kubenswrapper[4969]: I1004 08:36:49.946475 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:36:50 crc kubenswrapper[4969]: I1004 08:36:50.413113 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:36:50 crc kubenswrapper[4969]: I1004 08:36:50.540957 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerStarted","Data":"13ae23401025f1c7e27c81052bff279d623ad948115ac8ca0af2652325a8d006"} Oct 04 08:36:50 crc kubenswrapper[4969]: I1004 08:36:50.546304 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"434f8bca-e752-49f3-9bce-286c90b64c00","Type":"ContainerStarted","Data":"d58f4b2f4c1ff791a8169f57bed959083d777a86f23d646e8f4f2a238e5e31cc"} Oct 04 08:36:50 crc kubenswrapper[4969]: I1004 08:36:50.763317 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 08:36:51 crc kubenswrapper[4969]: I1004 08:36:51.067893 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe" path="/var/lib/kubelet/pods/25d0fa0e-2b33-49b3-b1a8-a95f1e9be6fe/volumes" Oct 04 08:36:51 crc kubenswrapper[4969]: I1004 08:36:51.564988 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"434f8bca-e752-49f3-9bce-286c90b64c00","Type":"ContainerStarted","Data":"719229648b3402f7be8b44f28d5e7c0fdc10c969726fb25eaa8234391f8e0d0b"} Oct 04 08:36:51 crc kubenswrapper[4969]: I1004 08:36:51.583073 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.583049407 podStartE2EDuration="2.583049407s" podCreationTimestamp="2025-10-04 08:36:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:36:51.580299986 +0000 UTC m=+1239.334568820" watchObservedRunningTime="2025-10-04 08:36:51.583049407 +0000 UTC m=+1239.337318261" Oct 04 08:36:52 crc kubenswrapper[4969]: I1004 08:36:52.578966 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerStarted","Data":"b10977563312793ba1ddd5c043d43a0e09269b8825d400f40a0dca08696f9d45"} Oct 04 08:36:52 crc kubenswrapper[4969]: I1004 08:36:52.607527 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.852905919 podStartE2EDuration="5.607510494s" podCreationTimestamp="2025-10-04 08:36:47 +0000 UTC" firstStartedPulling="2025-10-04 08:36:48.508639328 +0000 UTC m=+1236.262908142" lastFinishedPulling="2025-10-04 08:36:51.263243903 +0000 UTC m=+1239.017512717" observedRunningTime="2025-10-04 08:36:52.603337586 +0000 UTC m=+1240.357606430" watchObservedRunningTime="2025-10-04 08:36:52.607510494 +0000 UTC m=+1240.361779308" Oct 04 08:36:53 crc kubenswrapper[4969]: I1004 08:36:53.588188 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 08:36:54 crc kubenswrapper[4969]: I1004 08:36:54.823282 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 08:36:54 crc kubenswrapper[4969]: I1004 08:36:54.823643 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 08:36:54 crc kubenswrapper[4969]: I1004 08:36:54.946730 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 08:36:55 crc kubenswrapper[4969]: I1004 08:36:55.844575 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 08:36:55 crc kubenswrapper[4969]: I1004 08:36:55.844594 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 08:36:56 crc kubenswrapper[4969]: I1004 08:36:56.893867 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 08:36:56 crc kubenswrapper[4969]: I1004 08:36:56.894239 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 08:36:57 crc kubenswrapper[4969]: I1004 08:36:57.936879 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 08:36:57 crc kubenswrapper[4969]: I1004 08:36:57.977719 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 08:36:57 crc kubenswrapper[4969]: I1004 08:36:57.977743 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 08:36:59 crc kubenswrapper[4969]: I1004 08:36:59.947195 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 08:36:59 crc kubenswrapper[4969]: I1004 08:36:59.985798 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 08:37:00 crc kubenswrapper[4969]: I1004 08:37:00.731995 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 08:37:04 crc kubenswrapper[4969]: I1004 08:37:04.829221 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 08:37:04 crc kubenswrapper[4969]: I1004 08:37:04.831396 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 08:37:04 crc kubenswrapper[4969]: I1004 08:37:04.836211 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 08:37:04 crc kubenswrapper[4969]: I1004 08:37:04.836359 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.743724 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.744790 4969 generic.go:334] "Generic (PLEG): container finished" podID="6dfd83e4-c050-45a4-905d-331a3aff0f9e" containerID="8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72" exitCode=137 Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.745334 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6dfd83e4-c050-45a4-905d-331a3aff0f9e","Type":"ContainerDied","Data":"8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72"} Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.745382 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6dfd83e4-c050-45a4-905d-331a3aff0f9e","Type":"ContainerDied","Data":"0fdaad971a78db81e70f465a5152b637eec9933d182ed9ebe6b2b22242ceb471"} Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.745399 4969 scope.go:117] "RemoveContainer" containerID="8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.782010 4969 scope.go:117] "RemoveContainer" containerID="8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72" Oct 04 08:37:06 crc kubenswrapper[4969]: E1004 08:37:06.782547 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72\": container with ID starting with 8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72 not found: ID does not exist" containerID="8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.782601 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72"} err="failed to get container status \"8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72\": rpc error: code = NotFound desc = could not find container \"8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72\": container with ID starting with 8fda020675c7aebac289c672f0ea3c4c1e0cf108af6b32b01eaf4ce0f14f2b72 not found: ID does not exist" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.903066 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.903770 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.905953 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.910714 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.930135 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnlmd\" (UniqueName: \"kubernetes.io/projected/6dfd83e4-c050-45a4-905d-331a3aff0f9e-kube-api-access-vnlmd\") pod \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.930364 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-combined-ca-bundle\") pod \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.930537 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-config-data\") pod \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\" (UID: \"6dfd83e4-c050-45a4-905d-331a3aff0f9e\") " Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.939745 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dfd83e4-c050-45a4-905d-331a3aff0f9e-kube-api-access-vnlmd" (OuterVolumeSpecName: "kube-api-access-vnlmd") pod "6dfd83e4-c050-45a4-905d-331a3aff0f9e" (UID: "6dfd83e4-c050-45a4-905d-331a3aff0f9e"). InnerVolumeSpecName "kube-api-access-vnlmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.968886 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6dfd83e4-c050-45a4-905d-331a3aff0f9e" (UID: "6dfd83e4-c050-45a4-905d-331a3aff0f9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:06 crc kubenswrapper[4969]: I1004 08:37:06.995681 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-config-data" (OuterVolumeSpecName: "config-data") pod "6dfd83e4-c050-45a4-905d-331a3aff0f9e" (UID: "6dfd83e4-c050-45a4-905d-331a3aff0f9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.032924 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnlmd\" (UniqueName: \"kubernetes.io/projected/6dfd83e4-c050-45a4-905d-331a3aff0f9e-kube-api-access-vnlmd\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.032957 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.033100 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfd83e4-c050-45a4-905d-331a3aff0f9e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.760630 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.761711 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.775708 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.796856 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.820235 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.856578 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 08:37:07 crc kubenswrapper[4969]: E1004 08:37:07.857737 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dfd83e4-c050-45a4-905d-331a3aff0f9e" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.857768 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dfd83e4-c050-45a4-905d-331a3aff0f9e" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.858271 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dfd83e4-c050-45a4-905d-331a3aff0f9e" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.859560 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.862552 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.862591 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.862916 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.908488 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.967099 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.967167 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-855s8\" (UniqueName: \"kubernetes.io/projected/2075dd48-ec13-4ebb-841b-3a2d82724f98-kube-api-access-855s8\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.967353 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.967611 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.967674 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:07 crc kubenswrapper[4969]: I1004 08:37:07.996634 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6df8b76f9c-knxjd"] Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.003700 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.011315 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6df8b76f9c-knxjd"] Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.069538 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.069600 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-855s8\" (UniqueName: \"kubernetes.io/projected/2075dd48-ec13-4ebb-841b-3a2d82724f98-kube-api-access-855s8\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.069627 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.069672 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.069694 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.074603 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.078047 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.084380 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.086622 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-855s8\" (UniqueName: \"kubernetes.io/projected/2075dd48-ec13-4ebb-841b-3a2d82724f98-kube-api-access-855s8\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.105578 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/2075dd48-ec13-4ebb-841b-3a2d82724f98-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2075dd48-ec13-4ebb-841b-3a2d82724f98\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.172352 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-sb\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.172397 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-swift-storage-0\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.172414 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnx7b\" (UniqueName: \"kubernetes.io/projected/ddb0ee51-df73-4216-ad98-2efcf1529efa-kube-api-access-dnx7b\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.172484 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-nb\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.172550 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-svc\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.172592 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-config\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.201084 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.274497 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-sb\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.274553 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-swift-storage-0\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.274573 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnx7b\" (UniqueName: \"kubernetes.io/projected/ddb0ee51-df73-4216-ad98-2efcf1529efa-kube-api-access-dnx7b\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.274592 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-nb\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.274627 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-svc\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.274669 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-config\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.275613 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-swift-storage-0\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.275686 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-sb\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.275714 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-nb\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.275760 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-config\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.275852 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-svc\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.296575 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnx7b\" (UniqueName: \"kubernetes.io/projected/ddb0ee51-df73-4216-ad98-2efcf1529efa-kube-api-access-dnx7b\") pod \"dnsmasq-dns-6df8b76f9c-knxjd\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.331368 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.727003 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.772553 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2075dd48-ec13-4ebb-841b-3a2d82724f98","Type":"ContainerStarted","Data":"55d7f0f6165f51744c0cd399b2bdc8282ec12e2333f80e24b08a015d1da7896b"} Oct 04 08:37:08 crc kubenswrapper[4969]: I1004 08:37:08.822837 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6df8b76f9c-knxjd"] Oct 04 08:37:08 crc kubenswrapper[4969]: W1004 08:37:08.840279 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddb0ee51_df73_4216_ad98_2efcf1529efa.slice/crio-f48e37fefe04c31ac1b9b82317a1dcc55cb74134ab09de3ec5303b80cd9c4d3e WatchSource:0}: Error finding container f48e37fefe04c31ac1b9b82317a1dcc55cb74134ab09de3ec5303b80cd9c4d3e: Status 404 returned error can't find the container with id f48e37fefe04c31ac1b9b82317a1dcc55cb74134ab09de3ec5303b80cd9c4d3e Oct 04 08:37:09 crc kubenswrapper[4969]: I1004 08:37:09.069957 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dfd83e4-c050-45a4-905d-331a3aff0f9e" path="/var/lib/kubelet/pods/6dfd83e4-c050-45a4-905d-331a3aff0f9e/volumes" Oct 04 08:37:09 crc kubenswrapper[4969]: I1004 08:37:09.785789 4969 generic.go:334] "Generic (PLEG): container finished" podID="ddb0ee51-df73-4216-ad98-2efcf1529efa" containerID="b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7" exitCode=0 Oct 04 08:37:09 crc kubenswrapper[4969]: I1004 08:37:09.785885 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" event={"ID":"ddb0ee51-df73-4216-ad98-2efcf1529efa","Type":"ContainerDied","Data":"b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7"} Oct 04 08:37:09 crc kubenswrapper[4969]: I1004 08:37:09.786125 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" event={"ID":"ddb0ee51-df73-4216-ad98-2efcf1529efa","Type":"ContainerStarted","Data":"f48e37fefe04c31ac1b9b82317a1dcc55cb74134ab09de3ec5303b80cd9c4d3e"} Oct 04 08:37:09 crc kubenswrapper[4969]: I1004 08:37:09.788432 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2075dd48-ec13-4ebb-841b-3a2d82724f98","Type":"ContainerStarted","Data":"b053f5bf8d6b7c51d253b80774f39375349b877caf79d1ad04b4129b7800e926"} Oct 04 08:37:09 crc kubenswrapper[4969]: I1004 08:37:09.874154 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.8741165889999998 podStartE2EDuration="2.874116589s" podCreationTimestamp="2025-10-04 08:37:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:37:09.866654191 +0000 UTC m=+1257.620923005" watchObservedRunningTime="2025-10-04 08:37:09.874116589 +0000 UTC m=+1257.628385403" Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.427782 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.475170 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.475477 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="ceilometer-central-agent" containerID="cri-o://8a7f7c9681d88161014fb95f54a27fed0435322438935efcda4191c686af1022" gracePeriod=30 Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.475625 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="proxy-httpd" containerID="cri-o://b10977563312793ba1ddd5c043d43a0e09269b8825d400f40a0dca08696f9d45" gracePeriod=30 Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.475668 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="sg-core" containerID="cri-o://13ae23401025f1c7e27c81052bff279d623ad948115ac8ca0af2652325a8d006" gracePeriod=30 Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.475704 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="ceilometer-notification-agent" containerID="cri-o://3436b85a700dab71576f2b84084f24b062c0d248728d5ca6871242ea7986d689" gracePeriod=30 Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.489170 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.216:3000/\": EOF" Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.798698 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" event={"ID":"ddb0ee51-df73-4216-ad98-2efcf1529efa","Type":"ContainerStarted","Data":"50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab"} Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.799035 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.801490 4969 generic.go:334] "Generic (PLEG): container finished" podID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerID="b10977563312793ba1ddd5c043d43a0e09269b8825d400f40a0dca08696f9d45" exitCode=0 Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.801520 4969 generic.go:334] "Generic (PLEG): container finished" podID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerID="13ae23401025f1c7e27c81052bff279d623ad948115ac8ca0af2652325a8d006" exitCode=2 Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.801673 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-log" containerID="cri-o://a55d3802ea7dc76dcd28478a0d3a7f6944d37e3892d2b4ae2983a5a0ad86c261" gracePeriod=30 Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.801866 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerDied","Data":"b10977563312793ba1ddd5c043d43a0e09269b8825d400f40a0dca08696f9d45"} Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.801894 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerDied","Data":"13ae23401025f1c7e27c81052bff279d623ad948115ac8ca0af2652325a8d006"} Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.802395 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-api" containerID="cri-o://619b585ea6b3e97697dda5987967b64d34cc5bb22638bf518754b40debef639f" gracePeriod=30 Oct 04 08:37:10 crc kubenswrapper[4969]: I1004 08:37:10.832294 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" podStartSLOduration=3.832275306 podStartE2EDuration="3.832275306s" podCreationTimestamp="2025-10-04 08:37:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:37:10.821922415 +0000 UTC m=+1258.576191229" watchObservedRunningTime="2025-10-04 08:37:10.832275306 +0000 UTC m=+1258.586544120" Oct 04 08:37:11 crc kubenswrapper[4969]: I1004 08:37:11.813858 4969 generic.go:334] "Generic (PLEG): container finished" podID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerID="8a7f7c9681d88161014fb95f54a27fed0435322438935efcda4191c686af1022" exitCode=0 Oct 04 08:37:11 crc kubenswrapper[4969]: I1004 08:37:11.813933 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerDied","Data":"8a7f7c9681d88161014fb95f54a27fed0435322438935efcda4191c686af1022"} Oct 04 08:37:11 crc kubenswrapper[4969]: I1004 08:37:11.816170 4969 generic.go:334] "Generic (PLEG): container finished" podID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerID="a55d3802ea7dc76dcd28478a0d3a7f6944d37e3892d2b4ae2983a5a0ad86c261" exitCode=143 Oct 04 08:37:11 crc kubenswrapper[4969]: I1004 08:37:11.816252 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"451dbcc5-7ad5-429a-8404-73b2c75466bd","Type":"ContainerDied","Data":"a55d3802ea7dc76dcd28478a0d3a7f6944d37e3892d2b4ae2983a5a0ad86c261"} Oct 04 08:37:12 crc kubenswrapper[4969]: I1004 08:37:12.827820 4969 generic.go:334] "Generic (PLEG): container finished" podID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerID="619b585ea6b3e97697dda5987967b64d34cc5bb22638bf518754b40debef639f" exitCode=0 Oct 04 08:37:12 crc kubenswrapper[4969]: I1004 08:37:12.827867 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"451dbcc5-7ad5-429a-8404-73b2c75466bd","Type":"ContainerDied","Data":"619b585ea6b3e97697dda5987967b64d34cc5bb22638bf518754b40debef639f"} Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.025575 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.192825 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-combined-ca-bundle\") pod \"451dbcc5-7ad5-429a-8404-73b2c75466bd\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.193617 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/451dbcc5-7ad5-429a-8404-73b2c75466bd-logs\") pod \"451dbcc5-7ad5-429a-8404-73b2c75466bd\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.193803 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vzgb\" (UniqueName: \"kubernetes.io/projected/451dbcc5-7ad5-429a-8404-73b2c75466bd-kube-api-access-2vzgb\") pod \"451dbcc5-7ad5-429a-8404-73b2c75466bd\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.193917 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-config-data\") pod \"451dbcc5-7ad5-429a-8404-73b2c75466bd\" (UID: \"451dbcc5-7ad5-429a-8404-73b2c75466bd\") " Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.194526 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/451dbcc5-7ad5-429a-8404-73b2c75466bd-logs" (OuterVolumeSpecName: "logs") pod "451dbcc5-7ad5-429a-8404-73b2c75466bd" (UID: "451dbcc5-7ad5-429a-8404-73b2c75466bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.194944 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/451dbcc5-7ad5-429a-8404-73b2c75466bd-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.202992 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.216588 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/451dbcc5-7ad5-429a-8404-73b2c75466bd-kube-api-access-2vzgb" (OuterVolumeSpecName: "kube-api-access-2vzgb") pod "451dbcc5-7ad5-429a-8404-73b2c75466bd" (UID: "451dbcc5-7ad5-429a-8404-73b2c75466bd"). InnerVolumeSpecName "kube-api-access-2vzgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.234364 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "451dbcc5-7ad5-429a-8404-73b2c75466bd" (UID: "451dbcc5-7ad5-429a-8404-73b2c75466bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.247602 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-config-data" (OuterVolumeSpecName: "config-data") pod "451dbcc5-7ad5-429a-8404-73b2c75466bd" (UID: "451dbcc5-7ad5-429a-8404-73b2c75466bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.297189 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.297219 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vzgb\" (UniqueName: \"kubernetes.io/projected/451dbcc5-7ad5-429a-8404-73b2c75466bd-kube-api-access-2vzgb\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.297229 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/451dbcc5-7ad5-429a-8404-73b2c75466bd-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.840849 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"451dbcc5-7ad5-429a-8404-73b2c75466bd","Type":"ContainerDied","Data":"485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f"} Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.842149 4969 scope.go:117] "RemoveContainer" containerID="619b585ea6b3e97697dda5987967b64d34cc5bb22638bf518754b40debef639f" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.840906 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.878388 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.895393 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.895798 4969 scope.go:117] "RemoveContainer" containerID="a55d3802ea7dc76dcd28478a0d3a7f6944d37e3892d2b4ae2983a5a0ad86c261" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.906555 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:13 crc kubenswrapper[4969]: E1004 08:37:13.906991 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-api" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.907009 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-api" Oct 04 08:37:13 crc kubenswrapper[4969]: E1004 08:37:13.907041 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-log" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.907048 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-log" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.907241 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-api" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.907270 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" containerName="nova-api-log" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.908275 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.911731 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.911793 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.912001 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 08:37:13 crc kubenswrapper[4969]: I1004 08:37:13.931055 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.009732 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-public-tls-certs\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.009800 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.009834 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.009860 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfl94\" (UniqueName: \"kubernetes.io/projected/4f441be3-f48c-4e99-8506-eb63c474600f-kube-api-access-sfl94\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.009967 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f441be3-f48c-4e99-8506-eb63c474600f-logs\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.009992 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-config-data\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.112105 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-public-tls-certs\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.112211 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.112262 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.112303 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfl94\" (UniqueName: \"kubernetes.io/projected/4f441be3-f48c-4e99-8506-eb63c474600f-kube-api-access-sfl94\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.112408 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f441be3-f48c-4e99-8506-eb63c474600f-logs\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.112465 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-config-data\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.114090 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f441be3-f48c-4e99-8506-eb63c474600f-logs\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.117131 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.118361 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-public-tls-certs\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.120745 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-config-data\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.121892 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.133641 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfl94\" (UniqueName: \"kubernetes.io/projected/4f441be3-f48c-4e99-8506-eb63c474600f-kube-api-access-sfl94\") pod \"nova-api-0\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.237294 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.719974 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:14 crc kubenswrapper[4969]: I1004 08:37:14.851069 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f441be3-f48c-4e99-8506-eb63c474600f","Type":"ContainerStarted","Data":"7e413aab11c865c2c3d89705e6f6b31bf5744727a953513cf2d3fede4737aa4b"} Oct 04 08:37:15 crc kubenswrapper[4969]: I1004 08:37:15.075918 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="451dbcc5-7ad5-429a-8404-73b2c75466bd" path="/var/lib/kubelet/pods/451dbcc5-7ad5-429a-8404-73b2c75466bd/volumes" Oct 04 08:37:15 crc kubenswrapper[4969]: I1004 08:37:15.868280 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f441be3-f48c-4e99-8506-eb63c474600f","Type":"ContainerStarted","Data":"e98748457542f67932e7a881b391031fd1270c8121f5d1322bc18d2b90d99c0c"} Oct 04 08:37:15 crc kubenswrapper[4969]: I1004 08:37:15.868654 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f441be3-f48c-4e99-8506-eb63c474600f","Type":"ContainerStarted","Data":"48e54196ea4881f5920959b82150bebf4c59603fa0d9fc348cd424d36838b7c7"} Oct 04 08:37:15 crc kubenswrapper[4969]: I1004 08:37:15.917770 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.917749257 podStartE2EDuration="2.917749257s" podCreationTimestamp="2025-10-04 08:37:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:37:15.898560191 +0000 UTC m=+1263.652829055" watchObservedRunningTime="2025-10-04 08:37:15.917749257 +0000 UTC m=+1263.672018081" Oct 04 08:37:16 crc kubenswrapper[4969]: E1004 08:37:16.723396 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod451dbcc5_7ad5_429a_8404_73b2c75466bd.slice/crio-485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f\": RecentStats: unable to find data in memory cache]" Oct 04 08:37:16 crc kubenswrapper[4969]: I1004 08:37:16.884248 4969 generic.go:334] "Generic (PLEG): container finished" podID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerID="3436b85a700dab71576f2b84084f24b062c0d248728d5ca6871242ea7986d689" exitCode=0 Oct 04 08:37:16 crc kubenswrapper[4969]: I1004 08:37:16.884365 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerDied","Data":"3436b85a700dab71576f2b84084f24b062c0d248728d5ca6871242ea7986d689"} Oct 04 08:37:16 crc kubenswrapper[4969]: I1004 08:37:16.884438 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb36ec4c-072a-40b9-bd4d-a23938c71a67","Type":"ContainerDied","Data":"f2fe71499549387275bf3dc8d66d2c646e699d14ba088e11c401c16ffeed89df"} Oct 04 08:37:16 crc kubenswrapper[4969]: I1004 08:37:16.884453 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2fe71499549387275bf3dc8d66d2c646e699d14ba088e11c401c16ffeed89df" Oct 04 08:37:16 crc kubenswrapper[4969]: I1004 08:37:16.961460 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.023252 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-ceilometer-tls-certs\") pod \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.023295 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-scripts\") pod \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.023320 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-config-data\") pod \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.023352 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-run-httpd\") pod \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.023428 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-combined-ca-bundle\") pod \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.023487 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-sg-core-conf-yaml\") pod \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.023516 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7xsm\" (UniqueName: \"kubernetes.io/projected/bb36ec4c-072a-40b9-bd4d-a23938c71a67-kube-api-access-z7xsm\") pod \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.023538 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-log-httpd\") pod \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\" (UID: \"bb36ec4c-072a-40b9-bd4d-a23938c71a67\") " Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.024125 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bb36ec4c-072a-40b9-bd4d-a23938c71a67" (UID: "bb36ec4c-072a-40b9-bd4d-a23938c71a67"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.024167 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bb36ec4c-072a-40b9-bd4d-a23938c71a67" (UID: "bb36ec4c-072a-40b9-bd4d-a23938c71a67"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.040673 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-scripts" (OuterVolumeSpecName: "scripts") pod "bb36ec4c-072a-40b9-bd4d-a23938c71a67" (UID: "bb36ec4c-072a-40b9-bd4d-a23938c71a67"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.105833 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb36ec4c-072a-40b9-bd4d-a23938c71a67-kube-api-access-z7xsm" (OuterVolumeSpecName: "kube-api-access-z7xsm") pod "bb36ec4c-072a-40b9-bd4d-a23938c71a67" (UID: "bb36ec4c-072a-40b9-bd4d-a23938c71a67"). InnerVolumeSpecName "kube-api-access-z7xsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.129686 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7xsm\" (UniqueName: \"kubernetes.io/projected/bb36ec4c-072a-40b9-bd4d-a23938c71a67-kube-api-access-z7xsm\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.129920 4969 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.129995 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.130051 4969 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb36ec4c-072a-40b9-bd4d-a23938c71a67-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.161562 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bb36ec4c-072a-40b9-bd4d-a23938c71a67" (UID: "bb36ec4c-072a-40b9-bd4d-a23938c71a67"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.192220 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bb36ec4c-072a-40b9-bd4d-a23938c71a67" (UID: "bb36ec4c-072a-40b9-bd4d-a23938c71a67"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.215940 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb36ec4c-072a-40b9-bd4d-a23938c71a67" (UID: "bb36ec4c-072a-40b9-bd4d-a23938c71a67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.237894 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-config-data" (OuterVolumeSpecName: "config-data") pod "bb36ec4c-072a-40b9-bd4d-a23938c71a67" (UID: "bb36ec4c-072a-40b9-bd4d-a23938c71a67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.248377 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.248830 4969 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.248863 4969 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.248873 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb36ec4c-072a-40b9-bd4d-a23938c71a67-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.896971 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.955757 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.978178 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.992686 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:37:17 crc kubenswrapper[4969]: E1004 08:37:17.993193 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="ceilometer-notification-agent" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.993213 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="ceilometer-notification-agent" Oct 04 08:37:17 crc kubenswrapper[4969]: E1004 08:37:17.993235 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="sg-core" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.993244 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="sg-core" Oct 04 08:37:17 crc kubenswrapper[4969]: E1004 08:37:17.993260 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="proxy-httpd" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.993270 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="proxy-httpd" Oct 04 08:37:17 crc kubenswrapper[4969]: E1004 08:37:17.993293 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="ceilometer-central-agent" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.993302 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="ceilometer-central-agent" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.993564 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="proxy-httpd" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.993590 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="sg-core" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.993608 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="ceilometer-notification-agent" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.993629 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" containerName="ceilometer-central-agent" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.995859 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.998746 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.998828 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 08:37:17 crc kubenswrapper[4969]: I1004 08:37:17.999231 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.001320 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.066894 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-scripts\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.067018 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-log-httpd\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.067052 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.067290 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-run-httpd\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.067355 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-config-data\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.067518 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs25g\" (UniqueName: \"kubernetes.io/projected/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-kube-api-access-xs25g\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.067596 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.067656 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.170232 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-scripts\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.170340 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-log-httpd\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.170362 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.170509 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-run-httpd\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.170551 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-config-data\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.170620 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs25g\" (UniqueName: \"kubernetes.io/projected/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-kube-api-access-xs25g\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.170659 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.170688 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.171099 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-log-httpd\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.171464 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-run-httpd\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.175133 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.175264 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-config-data\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.175538 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.176885 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-scripts\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.178848 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.188065 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs25g\" (UniqueName: \"kubernetes.io/projected/0a525e7b-7a1c-4671-a6cc-e81bd316fd48-kube-api-access-xs25g\") pod \"ceilometer-0\" (UID: \"0a525e7b-7a1c-4671-a6cc-e81bd316fd48\") " pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.201957 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.243405 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.333567 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.348743 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.436620 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6759cf46f9-qjbd6"] Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.436910 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" podUID="f764da96-9012-4a0a-9144-4922906ed2d3" containerName="dnsmasq-dns" containerID="cri-o://7fc94259ee054c59317135db33a8f1abdbf881dd61b963611392daed882ea91b" gracePeriod=10 Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.900872 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.919303 4969 generic.go:334] "Generic (PLEG): container finished" podID="f764da96-9012-4a0a-9144-4922906ed2d3" containerID="7fc94259ee054c59317135db33a8f1abdbf881dd61b963611392daed882ea91b" exitCode=0 Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.919733 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" event={"ID":"f764da96-9012-4a0a-9144-4922906ed2d3","Type":"ContainerDied","Data":"7fc94259ee054c59317135db33a8f1abdbf881dd61b963611392daed882ea91b"} Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.919777 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" event={"ID":"f764da96-9012-4a0a-9144-4922906ed2d3","Type":"ContainerDied","Data":"4350deba044ad72445a474dae4c995b511dfb8e2d4a0b7f1384d723bb498e963"} Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.919793 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4350deba044ad72445a474dae4c995b511dfb8e2d4a0b7f1384d723bb498e963" Oct 04 08:37:18 crc kubenswrapper[4969]: I1004 08:37:18.937592 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.028251 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.071856 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb36ec4c-072a-40b9-bd4d-a23938c71a67" path="/var/lib/kubelet/pods/bb36ec4c-072a-40b9-bd4d-a23938c71a67/volumes" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.096043 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-nb\") pod \"f764da96-9012-4a0a-9144-4922906ed2d3\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.096355 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-config\") pod \"f764da96-9012-4a0a-9144-4922906ed2d3\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.096637 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d8db\" (UniqueName: \"kubernetes.io/projected/f764da96-9012-4a0a-9144-4922906ed2d3-kube-api-access-4d8db\") pod \"f764da96-9012-4a0a-9144-4922906ed2d3\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.096782 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-swift-storage-0\") pod \"f764da96-9012-4a0a-9144-4922906ed2d3\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.096918 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-svc\") pod \"f764da96-9012-4a0a-9144-4922906ed2d3\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.097092 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-sb\") pod \"f764da96-9012-4a0a-9144-4922906ed2d3\" (UID: \"f764da96-9012-4a0a-9144-4922906ed2d3\") " Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.102241 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f764da96-9012-4a0a-9144-4922906ed2d3-kube-api-access-4d8db" (OuterVolumeSpecName: "kube-api-access-4d8db") pod "f764da96-9012-4a0a-9144-4922906ed2d3" (UID: "f764da96-9012-4a0a-9144-4922906ed2d3"). InnerVolumeSpecName "kube-api-access-4d8db". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.164198 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-tj6f9"] Oct 04 08:37:19 crc kubenswrapper[4969]: E1004 08:37:19.164910 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f764da96-9012-4a0a-9144-4922906ed2d3" containerName="init" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.164929 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f764da96-9012-4a0a-9144-4922906ed2d3" containerName="init" Oct 04 08:37:19 crc kubenswrapper[4969]: E1004 08:37:19.164937 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f764da96-9012-4a0a-9144-4922906ed2d3" containerName="dnsmasq-dns" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.164943 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f764da96-9012-4a0a-9144-4922906ed2d3" containerName="dnsmasq-dns" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.165139 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f764da96-9012-4a0a-9144-4922906ed2d3" containerName="dnsmasq-dns" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.165882 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.171030 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.171862 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.184285 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-tj6f9"] Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.206884 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f764da96-9012-4a0a-9144-4922906ed2d3" (UID: "f764da96-9012-4a0a-9144-4922906ed2d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.210691 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-scripts\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.210882 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8m2f\" (UniqueName: \"kubernetes.io/projected/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-kube-api-access-p8m2f\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.211169 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.211242 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-config-data\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.211371 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.211387 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d8db\" (UniqueName: \"kubernetes.io/projected/f764da96-9012-4a0a-9144-4922906ed2d3-kube-api-access-4d8db\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.212715 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f764da96-9012-4a0a-9144-4922906ed2d3" (UID: "f764da96-9012-4a0a-9144-4922906ed2d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.219404 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f764da96-9012-4a0a-9144-4922906ed2d3" (UID: "f764da96-9012-4a0a-9144-4922906ed2d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.233163 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-config" (OuterVolumeSpecName: "config") pod "f764da96-9012-4a0a-9144-4922906ed2d3" (UID: "f764da96-9012-4a0a-9144-4922906ed2d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.239309 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f764da96-9012-4a0a-9144-4922906ed2d3" (UID: "f764da96-9012-4a0a-9144-4922906ed2d3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.312912 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-scripts\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.313043 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8m2f\" (UniqueName: \"kubernetes.io/projected/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-kube-api-access-p8m2f\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.313155 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.313220 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-config-data\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.313352 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.313382 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.313401 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.313437 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f764da96-9012-4a0a-9144-4922906ed2d3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.316388 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-scripts\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.317004 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.317930 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-config-data\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.330862 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8m2f\" (UniqueName: \"kubernetes.io/projected/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-kube-api-access-p8m2f\") pod \"nova-cell1-cell-mapping-tj6f9\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.500222 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.931050 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a525e7b-7a1c-4671-a6cc-e81bd316fd48","Type":"ContainerStarted","Data":"1f6cde88b3ad84c4b8546597f144b19f6472a3495c1d8c5ef02f5fc8a814b611"} Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.931446 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a525e7b-7a1c-4671-a6cc-e81bd316fd48","Type":"ContainerStarted","Data":"55844fb58089e4b9ffac573acfa005875b27c7ae60162968ad9c061714b8d232"} Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.931463 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a525e7b-7a1c-4671-a6cc-e81bd316fd48","Type":"ContainerStarted","Data":"a7a062bdfd67d9610842a0a9d52608cfd02b3ecfc4d958d9abc343e4994c8bfa"} Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.931554 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6759cf46f9-qjbd6" Oct 04 08:37:19 crc kubenswrapper[4969]: I1004 08:37:19.985759 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6759cf46f9-qjbd6"] Oct 04 08:37:20 crc kubenswrapper[4969]: I1004 08:37:20.019468 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6759cf46f9-qjbd6"] Oct 04 08:37:20 crc kubenswrapper[4969]: I1004 08:37:20.053946 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-tj6f9"] Oct 04 08:37:20 crc kubenswrapper[4969]: I1004 08:37:20.943234 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a525e7b-7a1c-4671-a6cc-e81bd316fd48","Type":"ContainerStarted","Data":"5b9493ca3eb82f333ab032af7d27ed6e5a26e2dd454a61df212ab81444778fd6"} Oct 04 08:37:20 crc kubenswrapper[4969]: I1004 08:37:20.945679 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tj6f9" event={"ID":"1f7bb15e-5c1f-4811-acc0-421d2bd13d88","Type":"ContainerStarted","Data":"e7ff900bdfe9aa8996d2b493e7f1d534772119ba64f3a4f72ccf3a12f38e9ea7"} Oct 04 08:37:20 crc kubenswrapper[4969]: I1004 08:37:20.945737 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tj6f9" event={"ID":"1f7bb15e-5c1f-4811-acc0-421d2bd13d88","Type":"ContainerStarted","Data":"c18d6fbc5ec0b1dc4416c6d9a51b575d0ff71d5a3dce1363eed2ab1f72fc741c"} Oct 04 08:37:20 crc kubenswrapper[4969]: I1004 08:37:20.976576 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-tj6f9" podStartSLOduration=1.976554283 podStartE2EDuration="1.976554283s" podCreationTimestamp="2025-10-04 08:37:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:37:20.970329237 +0000 UTC m=+1268.724598051" watchObservedRunningTime="2025-10-04 08:37:20.976554283 +0000 UTC m=+1268.730823117" Oct 04 08:37:21 crc kubenswrapper[4969]: I1004 08:37:21.066968 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f764da96-9012-4a0a-9144-4922906ed2d3" path="/var/lib/kubelet/pods/f764da96-9012-4a0a-9144-4922906ed2d3/volumes" Oct 04 08:37:22 crc kubenswrapper[4969]: I1004 08:37:22.970959 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a525e7b-7a1c-4671-a6cc-e81bd316fd48","Type":"ContainerStarted","Data":"0cd4845c46f2fd30a3e1b5a07e6c7d6f70155d4104ad30618aa33ddbecf89e63"} Oct 04 08:37:22 crc kubenswrapper[4969]: I1004 08:37:22.973638 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 08:37:23 crc kubenswrapper[4969]: I1004 08:37:23.001277 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.185255875 podStartE2EDuration="6.001257907s" podCreationTimestamp="2025-10-04 08:37:17 +0000 UTC" firstStartedPulling="2025-10-04 08:37:18.913717232 +0000 UTC m=+1266.667986036" lastFinishedPulling="2025-10-04 08:37:21.729719224 +0000 UTC m=+1269.483988068" observedRunningTime="2025-10-04 08:37:22.993734117 +0000 UTC m=+1270.748002931" watchObservedRunningTime="2025-10-04 08:37:23.001257907 +0000 UTC m=+1270.755526721" Oct 04 08:37:24 crc kubenswrapper[4969]: I1004 08:37:24.239864 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 08:37:24 crc kubenswrapper[4969]: I1004 08:37:24.240220 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 08:37:25 crc kubenswrapper[4969]: I1004 08:37:25.252564 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 08:37:25 crc kubenswrapper[4969]: I1004 08:37:25.252619 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 08:37:26 crc kubenswrapper[4969]: I1004 08:37:26.005566 4969 generic.go:334] "Generic (PLEG): container finished" podID="1f7bb15e-5c1f-4811-acc0-421d2bd13d88" containerID="e7ff900bdfe9aa8996d2b493e7f1d534772119ba64f3a4f72ccf3a12f38e9ea7" exitCode=0 Oct 04 08:37:26 crc kubenswrapper[4969]: I1004 08:37:26.005620 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tj6f9" event={"ID":"1f7bb15e-5c1f-4811-acc0-421d2bd13d88","Type":"ContainerDied","Data":"e7ff900bdfe9aa8996d2b493e7f1d534772119ba64f3a4f72ccf3a12f38e9ea7"} Oct 04 08:37:27 crc kubenswrapper[4969]: E1004 08:37:27.042738 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod451dbcc5_7ad5_429a_8404_73b2c75466bd.slice/crio-485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f\": RecentStats: unable to find data in memory cache]" Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.475344 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.620282 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-combined-ca-bundle\") pod \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.620498 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-config-data\") pod \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.620636 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-scripts\") pod \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.620678 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8m2f\" (UniqueName: \"kubernetes.io/projected/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-kube-api-access-p8m2f\") pod \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\" (UID: \"1f7bb15e-5c1f-4811-acc0-421d2bd13d88\") " Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.626633 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-kube-api-access-p8m2f" (OuterVolumeSpecName: "kube-api-access-p8m2f") pod "1f7bb15e-5c1f-4811-acc0-421d2bd13d88" (UID: "1f7bb15e-5c1f-4811-acc0-421d2bd13d88"). InnerVolumeSpecName "kube-api-access-p8m2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.630092 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-scripts" (OuterVolumeSpecName: "scripts") pod "1f7bb15e-5c1f-4811-acc0-421d2bd13d88" (UID: "1f7bb15e-5c1f-4811-acc0-421d2bd13d88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.649872 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-config-data" (OuterVolumeSpecName: "config-data") pod "1f7bb15e-5c1f-4811-acc0-421d2bd13d88" (UID: "1f7bb15e-5c1f-4811-acc0-421d2bd13d88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.652549 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f7bb15e-5c1f-4811-acc0-421d2bd13d88" (UID: "1f7bb15e-5c1f-4811-acc0-421d2bd13d88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.725908 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.725952 4969 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.725965 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8m2f\" (UniqueName: \"kubernetes.io/projected/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-kube-api-access-p8m2f\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:27 crc kubenswrapper[4969]: I1004 08:37:27.725979 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f7bb15e-5c1f-4811-acc0-421d2bd13d88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.036031 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tj6f9" event={"ID":"1f7bb15e-5c1f-4811-acc0-421d2bd13d88","Type":"ContainerDied","Data":"c18d6fbc5ec0b1dc4416c6d9a51b575d0ff71d5a3dce1363eed2ab1f72fc741c"} Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.036088 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c18d6fbc5ec0b1dc4416c6d9a51b575d0ff71d5a3dce1363eed2ab1f72fc741c" Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.036177 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tj6f9" Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.261064 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.261684 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-log" containerID="cri-o://48e54196ea4881f5920959b82150bebf4c59603fa0d9fc348cd424d36838b7c7" gracePeriod=30 Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.262226 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-api" containerID="cri-o://e98748457542f67932e7a881b391031fd1270c8121f5d1322bc18d2b90d99c0c" gracePeriod=30 Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.310527 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.310797 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="434f8bca-e752-49f3-9bce-286c90b64c00" containerName="nova-scheduler-scheduler" containerID="cri-o://719229648b3402f7be8b44f28d5e7c0fdc10c969726fb25eaa8234391f8e0d0b" gracePeriod=30 Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.323741 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.323962 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-log" containerID="cri-o://99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500" gracePeriod=30 Oct 04 08:37:28 crc kubenswrapper[4969]: I1004 08:37:28.324509 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-metadata" containerID="cri-o://64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7" gracePeriod=30 Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.044841 4969 generic.go:334] "Generic (PLEG): container finished" podID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerID="99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500" exitCode=143 Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.045050 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"788a64ad-f6c3-46e5-941f-4b5f1fe568c7","Type":"ContainerDied","Data":"99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500"} Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.047312 4969 generic.go:334] "Generic (PLEG): container finished" podID="4f441be3-f48c-4e99-8506-eb63c474600f" containerID="48e54196ea4881f5920959b82150bebf4c59603fa0d9fc348cd424d36838b7c7" exitCode=143 Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.047395 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f441be3-f48c-4e99-8506-eb63c474600f","Type":"ContainerDied","Data":"48e54196ea4881f5920959b82150bebf4c59603fa0d9fc348cd424d36838b7c7"} Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.681762 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.766442 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-combined-ca-bundle\") pod \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.766544 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-nova-metadata-tls-certs\") pod \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.766573 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-logs\") pod \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.766742 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x722f\" (UniqueName: \"kubernetes.io/projected/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-kube-api-access-x722f\") pod \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.766773 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-config-data\") pod \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\" (UID: \"788a64ad-f6c3-46e5-941f-4b5f1fe568c7\") " Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.767170 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-logs" (OuterVolumeSpecName: "logs") pod "788a64ad-f6c3-46e5-941f-4b5f1fe568c7" (UID: "788a64ad-f6c3-46e5-941f-4b5f1fe568c7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.776590 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-kube-api-access-x722f" (OuterVolumeSpecName: "kube-api-access-x722f") pod "788a64ad-f6c3-46e5-941f-4b5f1fe568c7" (UID: "788a64ad-f6c3-46e5-941f-4b5f1fe568c7"). InnerVolumeSpecName "kube-api-access-x722f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.804027 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-config-data" (OuterVolumeSpecName: "config-data") pod "788a64ad-f6c3-46e5-941f-4b5f1fe568c7" (UID: "788a64ad-f6c3-46e5-941f-4b5f1fe568c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.806835 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "788a64ad-f6c3-46e5-941f-4b5f1fe568c7" (UID: "788a64ad-f6c3-46e5-941f-4b5f1fe568c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.844707 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "788a64ad-f6c3-46e5-941f-4b5f1fe568c7" (UID: "788a64ad-f6c3-46e5-941f-4b5f1fe568c7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.868375 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.868402 4969 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.868413 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.868433 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x722f\" (UniqueName: \"kubernetes.io/projected/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-kube-api-access-x722f\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:29 crc kubenswrapper[4969]: I1004 08:37:29.868443 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788a64ad-f6c3-46e5-941f-4b5f1fe568c7-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:29 crc kubenswrapper[4969]: E1004 08:37:29.949587 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="719229648b3402f7be8b44f28d5e7c0fdc10c969726fb25eaa8234391f8e0d0b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 08:37:29 crc kubenswrapper[4969]: E1004 08:37:29.951085 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="719229648b3402f7be8b44f28d5e7c0fdc10c969726fb25eaa8234391f8e0d0b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 08:37:29 crc kubenswrapper[4969]: E1004 08:37:29.952755 4969 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="719229648b3402f7be8b44f28d5e7c0fdc10c969726fb25eaa8234391f8e0d0b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 08:37:29 crc kubenswrapper[4969]: E1004 08:37:29.952795 4969 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="434f8bca-e752-49f3-9bce-286c90b64c00" containerName="nova-scheduler-scheduler" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.063511 4969 generic.go:334] "Generic (PLEG): container finished" podID="4f441be3-f48c-4e99-8506-eb63c474600f" containerID="e98748457542f67932e7a881b391031fd1270c8121f5d1322bc18d2b90d99c0c" exitCode=0 Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.063670 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f441be3-f48c-4e99-8506-eb63c474600f","Type":"ContainerDied","Data":"e98748457542f67932e7a881b391031fd1270c8121f5d1322bc18d2b90d99c0c"} Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.066291 4969 generic.go:334] "Generic (PLEG): container finished" podID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerID="64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7" exitCode=0 Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.066375 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.066382 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"788a64ad-f6c3-46e5-941f-4b5f1fe568c7","Type":"ContainerDied","Data":"64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7"} Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.066604 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"788a64ad-f6c3-46e5-941f-4b5f1fe568c7","Type":"ContainerDied","Data":"668f0417bff4817822ffe4c4ff8f063d04728ea8bc0b09ebd904cabad28b6e45"} Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.066653 4969 scope.go:117] "RemoveContainer" containerID="64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.129270 4969 scope.go:117] "RemoveContainer" containerID="99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.154498 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.168496 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.186408 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:37:30 crc kubenswrapper[4969]: E1004 08:37:30.186845 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-log" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.186860 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-log" Oct 04 08:37:30 crc kubenswrapper[4969]: E1004 08:37:30.186893 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7bb15e-5c1f-4811-acc0-421d2bd13d88" containerName="nova-manage" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.186900 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7bb15e-5c1f-4811-acc0-421d2bd13d88" containerName="nova-manage" Oct 04 08:37:30 crc kubenswrapper[4969]: E1004 08:37:30.186912 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-metadata" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.186918 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-metadata" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.187093 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-metadata" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.187106 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" containerName="nova-metadata-log" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.187117 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f7bb15e-5c1f-4811-acc0-421d2bd13d88" containerName="nova-manage" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.188108 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.191571 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.191731 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.196168 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.223723 4969 scope.go:117] "RemoveContainer" containerID="64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7" Oct 04 08:37:30 crc kubenswrapper[4969]: E1004 08:37:30.242629 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7\": container with ID starting with 64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7 not found: ID does not exist" containerID="64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.242679 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7"} err="failed to get container status \"64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7\": rpc error: code = NotFound desc = could not find container \"64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7\": container with ID starting with 64897b36e80c6df49c4ac1927ef8003c1acc889d88826f592d37ed6c3d512ac7 not found: ID does not exist" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.242707 4969 scope.go:117] "RemoveContainer" containerID="99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500" Oct 04 08:37:30 crc kubenswrapper[4969]: E1004 08:37:30.245572 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500\": container with ID starting with 99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500 not found: ID does not exist" containerID="99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.245597 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500"} err="failed to get container status \"99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500\": rpc error: code = NotFound desc = could not find container \"99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500\": container with ID starting with 99cdc00b666f922dc24e6ae972c8f8dab407ea532a15d4fca71ff457df29a500 not found: ID does not exist" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.275727 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7574220-81ab-4bde-b17b-c15d3339bfd6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.275798 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7574220-81ab-4bde-b17b-c15d3339bfd6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.275834 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7574220-81ab-4bde-b17b-c15d3339bfd6-logs\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.275864 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7574220-81ab-4bde-b17b-c15d3339bfd6-config-data\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.276364 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v27xr\" (UniqueName: \"kubernetes.io/projected/b7574220-81ab-4bde-b17b-c15d3339bfd6-kube-api-access-v27xr\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.349346 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.378573 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v27xr\" (UniqueName: \"kubernetes.io/projected/b7574220-81ab-4bde-b17b-c15d3339bfd6-kube-api-access-v27xr\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.378635 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7574220-81ab-4bde-b17b-c15d3339bfd6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.378662 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7574220-81ab-4bde-b17b-c15d3339bfd6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.378683 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7574220-81ab-4bde-b17b-c15d3339bfd6-logs\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.378702 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7574220-81ab-4bde-b17b-c15d3339bfd6-config-data\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.379938 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7574220-81ab-4bde-b17b-c15d3339bfd6-logs\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.384197 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7574220-81ab-4bde-b17b-c15d3339bfd6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.384735 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7574220-81ab-4bde-b17b-c15d3339bfd6-config-data\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.385547 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7574220-81ab-4bde-b17b-c15d3339bfd6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.396976 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v27xr\" (UniqueName: \"kubernetes.io/projected/b7574220-81ab-4bde-b17b-c15d3339bfd6-kube-api-access-v27xr\") pod \"nova-metadata-0\" (UID: \"b7574220-81ab-4bde-b17b-c15d3339bfd6\") " pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.480355 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfl94\" (UniqueName: \"kubernetes.io/projected/4f441be3-f48c-4e99-8506-eb63c474600f-kube-api-access-sfl94\") pod \"4f441be3-f48c-4e99-8506-eb63c474600f\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.480465 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-config-data\") pod \"4f441be3-f48c-4e99-8506-eb63c474600f\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.480558 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f441be3-f48c-4e99-8506-eb63c474600f-logs\") pod \"4f441be3-f48c-4e99-8506-eb63c474600f\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.480609 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-combined-ca-bundle\") pod \"4f441be3-f48c-4e99-8506-eb63c474600f\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.480779 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-public-tls-certs\") pod \"4f441be3-f48c-4e99-8506-eb63c474600f\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.480919 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-internal-tls-certs\") pod \"4f441be3-f48c-4e99-8506-eb63c474600f\" (UID: \"4f441be3-f48c-4e99-8506-eb63c474600f\") " Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.481819 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f441be3-f48c-4e99-8506-eb63c474600f-logs" (OuterVolumeSpecName: "logs") pod "4f441be3-f48c-4e99-8506-eb63c474600f" (UID: "4f441be3-f48c-4e99-8506-eb63c474600f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.485795 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f441be3-f48c-4e99-8506-eb63c474600f-kube-api-access-sfl94" (OuterVolumeSpecName: "kube-api-access-sfl94") pod "4f441be3-f48c-4e99-8506-eb63c474600f" (UID: "4f441be3-f48c-4e99-8506-eb63c474600f"). InnerVolumeSpecName "kube-api-access-sfl94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.514620 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-config-data" (OuterVolumeSpecName: "config-data") pod "4f441be3-f48c-4e99-8506-eb63c474600f" (UID: "4f441be3-f48c-4e99-8506-eb63c474600f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.520963 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.535524 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f441be3-f48c-4e99-8506-eb63c474600f" (UID: "4f441be3-f48c-4e99-8506-eb63c474600f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.552550 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4f441be3-f48c-4e99-8506-eb63c474600f" (UID: "4f441be3-f48c-4e99-8506-eb63c474600f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.558973 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4f441be3-f48c-4e99-8506-eb63c474600f" (UID: "4f441be3-f48c-4e99-8506-eb63c474600f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.592385 4969 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.592615 4969 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.592726 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfl94\" (UniqueName: \"kubernetes.io/projected/4f441be3-f48c-4e99-8506-eb63c474600f-kube-api-access-sfl94\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.592801 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.592858 4969 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f441be3-f48c-4e99-8506-eb63c474600f-logs\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:30 crc kubenswrapper[4969]: I1004 08:37:30.592912 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f441be3-f48c-4e99-8506-eb63c474600f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.009761 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 08:37:31 crc kubenswrapper[4969]: W1004 08:37:31.010869 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7574220_81ab_4bde_b17b_c15d3339bfd6.slice/crio-29c86a2d06fda3ec01056845843d8a515ee2caedd0de127ce64f125557656e19 WatchSource:0}: Error finding container 29c86a2d06fda3ec01056845843d8a515ee2caedd0de127ce64f125557656e19: Status 404 returned error can't find the container with id 29c86a2d06fda3ec01056845843d8a515ee2caedd0de127ce64f125557656e19 Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.074616 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788a64ad-f6c3-46e5-941f-4b5f1fe568c7" path="/var/lib/kubelet/pods/788a64ad-f6c3-46e5-941f-4b5f1fe568c7/volumes" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.091454 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f441be3-f48c-4e99-8506-eb63c474600f","Type":"ContainerDied","Data":"7e413aab11c865c2c3d89705e6f6b31bf5744727a953513cf2d3fede4737aa4b"} Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.091728 4969 scope.go:117] "RemoveContainer" containerID="e98748457542f67932e7a881b391031fd1270c8121f5d1322bc18d2b90d99c0c" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.091498 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.092739 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7574220-81ab-4bde-b17b-c15d3339bfd6","Type":"ContainerStarted","Data":"29c86a2d06fda3ec01056845843d8a515ee2caedd0de127ce64f125557656e19"} Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.131942 4969 scope.go:117] "RemoveContainer" containerID="48e54196ea4881f5920959b82150bebf4c59603fa0d9fc348cd424d36838b7c7" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.157233 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.176376 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.189743 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:31 crc kubenswrapper[4969]: E1004 08:37:31.190298 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-log" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.190321 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-log" Oct 04 08:37:31 crc kubenswrapper[4969]: E1004 08:37:31.190348 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-api" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.190356 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-api" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.190634 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-api" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.190667 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" containerName="nova-api-log" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.191939 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.195925 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.196500 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.196647 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.201694 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.317160 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-config-data\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.317429 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.317572 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-public-tls-certs\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.317634 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.317766 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5b33466-d882-4113-8c9a-bc1047ee6b9a-logs\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.317844 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8784k\" (UniqueName: \"kubernetes.io/projected/e5b33466-d882-4113-8c9a-bc1047ee6b9a-kube-api-access-8784k\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.419539 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-config-data\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.419627 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.419647 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-public-tls-certs\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.419719 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.419766 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5b33466-d882-4113-8c9a-bc1047ee6b9a-logs\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.419789 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8784k\" (UniqueName: \"kubernetes.io/projected/e5b33466-d882-4113-8c9a-bc1047ee6b9a-kube-api-access-8784k\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.420471 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5b33466-d882-4113-8c9a-bc1047ee6b9a-logs\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.425214 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.426043 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-config-data\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.426586 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.429207 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5b33466-d882-4113-8c9a-bc1047ee6b9a-public-tls-certs\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.444660 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8784k\" (UniqueName: \"kubernetes.io/projected/e5b33466-d882-4113-8c9a-bc1047ee6b9a-kube-api-access-8784k\") pod \"nova-api-0\" (UID: \"e5b33466-d882-4113-8c9a-bc1047ee6b9a\") " pod="openstack/nova-api-0" Oct 04 08:37:31 crc kubenswrapper[4969]: I1004 08:37:31.519732 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 08:37:32 crc kubenswrapper[4969]: W1004 08:37:31.983958 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5b33466_d882_4113_8c9a_bc1047ee6b9a.slice/crio-af930ff583d30a500729d0c0c8cf3be8be4d90fce52e9c047628afa41ae7c6b9 WatchSource:0}: Error finding container af930ff583d30a500729d0c0c8cf3be8be4d90fce52e9c047628afa41ae7c6b9: Status 404 returned error can't find the container with id af930ff583d30a500729d0c0c8cf3be8be4d90fce52e9c047628afa41ae7c6b9 Oct 04 08:37:32 crc kubenswrapper[4969]: I1004 08:37:31.987582 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 08:37:32 crc kubenswrapper[4969]: I1004 08:37:32.116504 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5b33466-d882-4113-8c9a-bc1047ee6b9a","Type":"ContainerStarted","Data":"af930ff583d30a500729d0c0c8cf3be8be4d90fce52e9c047628afa41ae7c6b9"} Oct 04 08:37:32 crc kubenswrapper[4969]: I1004 08:37:32.120322 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7574220-81ab-4bde-b17b-c15d3339bfd6","Type":"ContainerStarted","Data":"6e45ffc4e2d618211a0336ea94640ece98e8cf20182e338a04c0b949b1560b94"} Oct 04 08:37:32 crc kubenswrapper[4969]: I1004 08:37:32.120372 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7574220-81ab-4bde-b17b-c15d3339bfd6","Type":"ContainerStarted","Data":"1847edc77b211b3aef3acfa3a6772f6962153a7e9d2a9e21faf36094f00a86ef"} Oct 04 08:37:32 crc kubenswrapper[4969]: I1004 08:37:32.152228 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.152209622 podStartE2EDuration="2.152209622s" podCreationTimestamp="2025-10-04 08:37:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:37:32.147467785 +0000 UTC m=+1279.901736599" watchObservedRunningTime="2025-10-04 08:37:32.152209622 +0000 UTC m=+1279.906478436" Oct 04 08:37:33 crc kubenswrapper[4969]: I1004 08:37:33.068820 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f441be3-f48c-4e99-8506-eb63c474600f" path="/var/lib/kubelet/pods/4f441be3-f48c-4e99-8506-eb63c474600f/volumes" Oct 04 08:37:33 crc kubenswrapper[4969]: I1004 08:37:33.134383 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5b33466-d882-4113-8c9a-bc1047ee6b9a","Type":"ContainerStarted","Data":"e02034b06bf1ff57529d6d3c8b252ce4f72a500e983ced354026be7011ecfc09"} Oct 04 08:37:33 crc kubenswrapper[4969]: I1004 08:37:33.134498 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5b33466-d882-4113-8c9a-bc1047ee6b9a","Type":"ContainerStarted","Data":"cbb9cdf7543e71d3ec7c7f417213918296d47ad94753d1b6ae07af9d9578d986"} Oct 04 08:37:33 crc kubenswrapper[4969]: I1004 08:37:33.167840 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.167819885 podStartE2EDuration="2.167819885s" podCreationTimestamp="2025-10-04 08:37:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:37:33.156268562 +0000 UTC m=+1280.910537376" watchObservedRunningTime="2025-10-04 08:37:33.167819885 +0000 UTC m=+1280.922088699" Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.145494 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"434f8bca-e752-49f3-9bce-286c90b64c00","Type":"ContainerDied","Data":"719229648b3402f7be8b44f28d5e7c0fdc10c969726fb25eaa8234391f8e0d0b"} Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.145515 4969 generic.go:334] "Generic (PLEG): container finished" podID="434f8bca-e752-49f3-9bce-286c90b64c00" containerID="719229648b3402f7be8b44f28d5e7c0fdc10c969726fb25eaa8234391f8e0d0b" exitCode=0 Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.260570 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.387461 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-config-data\") pod \"434f8bca-e752-49f3-9bce-286c90b64c00\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.387639 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sthj\" (UniqueName: \"kubernetes.io/projected/434f8bca-e752-49f3-9bce-286c90b64c00-kube-api-access-4sthj\") pod \"434f8bca-e752-49f3-9bce-286c90b64c00\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.387817 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-combined-ca-bundle\") pod \"434f8bca-e752-49f3-9bce-286c90b64c00\" (UID: \"434f8bca-e752-49f3-9bce-286c90b64c00\") " Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.394341 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434f8bca-e752-49f3-9bce-286c90b64c00-kube-api-access-4sthj" (OuterVolumeSpecName: "kube-api-access-4sthj") pod "434f8bca-e752-49f3-9bce-286c90b64c00" (UID: "434f8bca-e752-49f3-9bce-286c90b64c00"). InnerVolumeSpecName "kube-api-access-4sthj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.422435 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "434f8bca-e752-49f3-9bce-286c90b64c00" (UID: "434f8bca-e752-49f3-9bce-286c90b64c00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.434042 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-config-data" (OuterVolumeSpecName: "config-data") pod "434f8bca-e752-49f3-9bce-286c90b64c00" (UID: "434f8bca-e752-49f3-9bce-286c90b64c00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.489857 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.489894 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sthj\" (UniqueName: \"kubernetes.io/projected/434f8bca-e752-49f3-9bce-286c90b64c00-kube-api-access-4sthj\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:34 crc kubenswrapper[4969]: I1004 08:37:34.489906 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f8bca-e752-49f3-9bce-286c90b64c00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.164161 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"434f8bca-e752-49f3-9bce-286c90b64c00","Type":"ContainerDied","Data":"d58f4b2f4c1ff791a8169f57bed959083d777a86f23d646e8f4f2a238e5e31cc"} Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.164256 4969 scope.go:117] "RemoveContainer" containerID="719229648b3402f7be8b44f28d5e7c0fdc10c969726fb25eaa8234391f8e0d0b" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.164473 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.210258 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.233020 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.249602 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:37:35 crc kubenswrapper[4969]: E1004 08:37:35.250397 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434f8bca-e752-49f3-9bce-286c90b64c00" containerName="nova-scheduler-scheduler" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.250475 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="434f8bca-e752-49f3-9bce-286c90b64c00" containerName="nova-scheduler-scheduler" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.251011 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="434f8bca-e752-49f3-9bce-286c90b64c00" containerName="nova-scheduler-scheduler" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.252602 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.255343 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.260385 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.409987 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcxhj\" (UniqueName: \"kubernetes.io/projected/9ef1ee4f-398a-49d8-a569-2d5bcc4cf073-kube-api-access-pcxhj\") pod \"nova-scheduler-0\" (UID: \"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073\") " pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.410217 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef1ee4f-398a-49d8-a569-2d5bcc4cf073-config-data\") pod \"nova-scheduler-0\" (UID: \"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073\") " pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.410269 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef1ee4f-398a-49d8-a569-2d5bcc4cf073-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073\") " pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.513202 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef1ee4f-398a-49d8-a569-2d5bcc4cf073-config-data\") pod \"nova-scheduler-0\" (UID: \"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073\") " pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.513312 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef1ee4f-398a-49d8-a569-2d5bcc4cf073-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073\") " pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.513662 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcxhj\" (UniqueName: \"kubernetes.io/projected/9ef1ee4f-398a-49d8-a569-2d5bcc4cf073-kube-api-access-pcxhj\") pod \"nova-scheduler-0\" (UID: \"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073\") " pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.521145 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.521195 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.521208 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef1ee4f-398a-49d8-a569-2d5bcc4cf073-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073\") " pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.531127 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef1ee4f-398a-49d8-a569-2d5bcc4cf073-config-data\") pod \"nova-scheduler-0\" (UID: \"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073\") " pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.547273 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcxhj\" (UniqueName: \"kubernetes.io/projected/9ef1ee4f-398a-49d8-a569-2d5bcc4cf073-kube-api-access-pcxhj\") pod \"nova-scheduler-0\" (UID: \"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073\") " pod="openstack/nova-scheduler-0" Oct 04 08:37:35 crc kubenswrapper[4969]: I1004 08:37:35.574346 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 08:37:36 crc kubenswrapper[4969]: I1004 08:37:36.076399 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 08:37:36 crc kubenswrapper[4969]: I1004 08:37:36.179503 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073","Type":"ContainerStarted","Data":"0e073124b94bc7ccc6934d86607c1ac699c1f4e7fcd3de43ca8d8ff7107a9e12"} Oct 04 08:37:37 crc kubenswrapper[4969]: I1004 08:37:37.073788 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434f8bca-e752-49f3-9bce-286c90b64c00" path="/var/lib/kubelet/pods/434f8bca-e752-49f3-9bce-286c90b64c00/volumes" Oct 04 08:37:37 crc kubenswrapper[4969]: I1004 08:37:37.206031 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9ef1ee4f-398a-49d8-a569-2d5bcc4cf073","Type":"ContainerStarted","Data":"88f076e4d269051d70e8b96a792fb387a46fe0f5c6787adba4584fe48b41f44b"} Oct 04 08:37:37 crc kubenswrapper[4969]: I1004 08:37:37.235686 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.235661735 podStartE2EDuration="2.235661735s" podCreationTimestamp="2025-10-04 08:37:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:37:37.230306086 +0000 UTC m=+1284.984574960" watchObservedRunningTime="2025-10-04 08:37:37.235661735 +0000 UTC m=+1284.989930559" Oct 04 08:37:37 crc kubenswrapper[4969]: E1004 08:37:37.395667 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod451dbcc5_7ad5_429a_8404_73b2c75466bd.slice/crio-485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f\": RecentStats: unable to find data in memory cache]" Oct 04 08:37:40 crc kubenswrapper[4969]: I1004 08:37:40.521528 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 08:37:40 crc kubenswrapper[4969]: I1004 08:37:40.521941 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 08:37:40 crc kubenswrapper[4969]: I1004 08:37:40.574727 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 08:37:41 crc kubenswrapper[4969]: I1004 08:37:41.520817 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 08:37:41 crc kubenswrapper[4969]: I1004 08:37:41.520886 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 08:37:41 crc kubenswrapper[4969]: I1004 08:37:41.542694 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b7574220-81ab-4bde-b17b-c15d3339bfd6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.223:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 08:37:41 crc kubenswrapper[4969]: I1004 08:37:41.542769 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b7574220-81ab-4bde-b17b-c15d3339bfd6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.223:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 08:37:42 crc kubenswrapper[4969]: I1004 08:37:42.602633 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5b33466-d882-4113-8c9a-bc1047ee6b9a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.224:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 08:37:42 crc kubenswrapper[4969]: I1004 08:37:42.602646 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5b33466-d882-4113-8c9a-bc1047ee6b9a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.224:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 08:37:45 crc kubenswrapper[4969]: I1004 08:37:45.574977 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 08:37:45 crc kubenswrapper[4969]: I1004 08:37:45.630355 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 08:37:46 crc kubenswrapper[4969]: I1004 08:37:46.392167 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 08:37:47 crc kubenswrapper[4969]: E1004 08:37:47.741006 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod451dbcc5_7ad5_429a_8404_73b2c75466bd.slice/crio-485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f\": RecentStats: unable to find data in memory cache]" Oct 04 08:37:48 crc kubenswrapper[4969]: I1004 08:37:48.442260 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 08:37:50 crc kubenswrapper[4969]: I1004 08:37:50.527984 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 08:37:50 crc kubenswrapper[4969]: I1004 08:37:50.534413 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 08:37:50 crc kubenswrapper[4969]: I1004 08:37:50.535725 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 08:37:51 crc kubenswrapper[4969]: I1004 08:37:51.394554 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 08:37:51 crc kubenswrapper[4969]: I1004 08:37:51.532252 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 08:37:51 crc kubenswrapper[4969]: I1004 08:37:51.532885 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 08:37:51 crc kubenswrapper[4969]: I1004 08:37:51.544598 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 08:37:51 crc kubenswrapper[4969]: I1004 08:37:51.562820 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 08:37:52 crc kubenswrapper[4969]: I1004 08:37:52.399285 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 08:37:52 crc kubenswrapper[4969]: I1004 08:37:52.413267 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 08:37:58 crc kubenswrapper[4969]: E1004 08:37:58.065287 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod451dbcc5_7ad5_429a_8404_73b2c75466bd.slice/crio-485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f\": RecentStats: unable to find data in memory cache]" Oct 04 08:38:00 crc kubenswrapper[4969]: I1004 08:38:00.735608 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 08:38:01 crc kubenswrapper[4969]: I1004 08:38:01.803969 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 08:38:04 crc kubenswrapper[4969]: I1004 08:38:04.192385 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1547b86a-ce65-4135-a8a5-957017c2271c" containerName="rabbitmq" containerID="cri-o://001e2338fea808e3665ffd16d9def28352878cb322c3ca5065371fdfd6537ced" gracePeriod=604797 Oct 04 08:38:05 crc kubenswrapper[4969]: I1004 08:38:05.213117 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="6eaebec4-b71b-409d-b91e-1993605ce209" containerName="rabbitmq" containerID="cri-o://eadc989bb8e449b5c0280b3c9f133dc07201cbfffbd1c533c6a2d8bae55469e2" gracePeriod=604797 Oct 04 08:38:05 crc kubenswrapper[4969]: I1004 08:38:05.860809 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1547b86a-ce65-4135-a8a5-957017c2271c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.201197 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="6eaebec4-b71b-409d-b91e-1993605ce209" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.604802 4969 generic.go:334] "Generic (PLEG): container finished" podID="1547b86a-ce65-4135-a8a5-957017c2271c" containerID="001e2338fea808e3665ffd16d9def28352878cb322c3ca5065371fdfd6537ced" exitCode=0 Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.605545 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1547b86a-ce65-4135-a8a5-957017c2271c","Type":"ContainerDied","Data":"001e2338fea808e3665ffd16d9def28352878cb322c3ca5065371fdfd6537ced"} Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.605603 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1547b86a-ce65-4135-a8a5-957017c2271c","Type":"ContainerDied","Data":"72cf44fdbaf1b0c34f3c10eca9aa9599f6773c839e258d471b1a9c4532e0689c"} Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.605622 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72cf44fdbaf1b0c34f3c10eca9aa9599f6773c839e258d471b1a9c4532e0689c" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.608724 4969 generic.go:334] "Generic (PLEG): container finished" podID="6eaebec4-b71b-409d-b91e-1993605ce209" containerID="eadc989bb8e449b5c0280b3c9f133dc07201cbfffbd1c533c6a2d8bae55469e2" exitCode=0 Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.608759 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6eaebec4-b71b-409d-b91e-1993605ce209","Type":"ContainerDied","Data":"eadc989bb8e449b5c0280b3c9f133dc07201cbfffbd1c533c6a2d8bae55469e2"} Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.655791 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.745467 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1547b86a-ce65-4135-a8a5-957017c2271c-erlang-cookie-secret\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.745627 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-tls\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.745662 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.745707 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7cmd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-kube-api-access-v7cmd\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.745798 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-plugins\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.745823 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-config-data\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.745884 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-plugins-conf\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.745927 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-erlang-cookie\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.746008 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1547b86a-ce65-4135-a8a5-957017c2271c-pod-info\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.746043 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-confd\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.746073 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-server-conf\") pod \"1547b86a-ce65-4135-a8a5-957017c2271c\" (UID: \"1547b86a-ce65-4135-a8a5-957017c2271c\") " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.747074 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.750613 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.754804 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.757341 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.760601 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1547b86a-ce65-4135-a8a5-957017c2271c-pod-info" (OuterVolumeSpecName: "pod-info") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.767752 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-kube-api-access-v7cmd" (OuterVolumeSpecName: "kube-api-access-v7cmd") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "kube-api-access-v7cmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.784647 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.798754 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1547b86a-ce65-4135-a8a5-957017c2271c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.835092 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-config-data" (OuterVolumeSpecName: "config-data") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.848479 4969 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1547b86a-ce65-4135-a8a5-957017c2271c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.848509 4969 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.848542 4969 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.848552 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7cmd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-kube-api-access-v7cmd\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.848563 4969 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.848572 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.848591 4969 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.848600 4969 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.848608 4969 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1547b86a-ce65-4135-a8a5-957017c2271c-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.885881 4969 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.936353 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-server-conf" (OuterVolumeSpecName: "server-conf") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.950018 4969 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.950049 4969 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1547b86a-ce65-4135-a8a5-957017c2271c-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.951832 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:06 crc kubenswrapper[4969]: I1004 08:38:06.979410 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1547b86a-ce65-4135-a8a5-957017c2271c" (UID: "1547b86a-ce65-4135-a8a5-957017c2271c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051006 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-config-data\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051068 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-tls\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051170 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eaebec4-b71b-409d-b91e-1993605ce209-erlang-cookie-secret\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051208 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eaebec4-b71b-409d-b91e-1993605ce209-pod-info\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051242 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-confd\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051285 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-plugins\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051340 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffm99\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-kube-api-access-ffm99\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051460 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-erlang-cookie\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051517 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-plugins-conf\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.051541 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-server-conf\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.052532 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"6eaebec4-b71b-409d-b91e-1993605ce209\" (UID: \"6eaebec4-b71b-409d-b91e-1993605ce209\") " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.052652 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.053008 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.055542 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eaebec4-b71b-409d-b91e-1993605ce209-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.055777 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.055804 4969 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eaebec4-b71b-409d-b91e-1993605ce209-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.055828 4969 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.055839 4969 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.055850 4969 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1547b86a-ce65-4135-a8a5-957017c2271c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.064438 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.064981 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.064990 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6eaebec4-b71b-409d-b91e-1993605ce209-pod-info" (OuterVolumeSpecName: "pod-info") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.065134 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-kube-api-access-ffm99" (OuterVolumeSpecName: "kube-api-access-ffm99") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "kube-api-access-ffm99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.082520 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-config-data" (OuterVolumeSpecName: "config-data") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.116973 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-server-conf" (OuterVolumeSpecName: "server-conf") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.157925 4969 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eaebec4-b71b-409d-b91e-1993605ce209-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.157960 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffm99\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-kube-api-access-ffm99\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.157970 4969 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.157978 4969 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.157998 4969 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.158006 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eaebec4-b71b-409d-b91e-1993605ce209-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.158014 4969 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.187351 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6eaebec4-b71b-409d-b91e-1993605ce209" (UID: "6eaebec4-b71b-409d-b91e-1993605ce209"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.206696 4969 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.259860 4969 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.259901 4969 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eaebec4-b71b-409d-b91e-1993605ce209-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.620653 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6eaebec4-b71b-409d-b91e-1993605ce209","Type":"ContainerDied","Data":"1e2b8014fdc03093198f115eb7098c67c5d21c3d5d9ba5d3b773ba5a903d006f"} Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.620691 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.620710 4969 scope.go:117] "RemoveContainer" containerID="eadc989bb8e449b5c0280b3c9f133dc07201cbfffbd1c533c6a2d8bae55469e2" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.620742 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.658173 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.674098 4969 scope.go:117] "RemoveContainer" containerID="8b1f66b1e50805e2ec1fadbeadbbdb658b9a97373b0469c6ef10e190672e5c1c" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.680061 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.693708 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.728838 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.742646 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 08:38:07 crc kubenswrapper[4969]: E1004 08:38:07.743041 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eaebec4-b71b-409d-b91e-1993605ce209" containerName="rabbitmq" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.743056 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eaebec4-b71b-409d-b91e-1993605ce209" containerName="rabbitmq" Oct 04 08:38:07 crc kubenswrapper[4969]: E1004 08:38:07.743083 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eaebec4-b71b-409d-b91e-1993605ce209" containerName="setup-container" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.743090 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eaebec4-b71b-409d-b91e-1993605ce209" containerName="setup-container" Oct 04 08:38:07 crc kubenswrapper[4969]: E1004 08:38:07.743103 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1547b86a-ce65-4135-a8a5-957017c2271c" containerName="rabbitmq" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.743108 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1547b86a-ce65-4135-a8a5-957017c2271c" containerName="rabbitmq" Oct 04 08:38:07 crc kubenswrapper[4969]: E1004 08:38:07.743131 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1547b86a-ce65-4135-a8a5-957017c2271c" containerName="setup-container" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.743136 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="1547b86a-ce65-4135-a8a5-957017c2271c" containerName="setup-container" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.743326 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eaebec4-b71b-409d-b91e-1993605ce209" containerName="rabbitmq" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.743342 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="1547b86a-ce65-4135-a8a5-957017c2271c" containerName="rabbitmq" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.744342 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.749974 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.749999 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.750043 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.750044 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.750259 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.750337 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-w9hfz" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.750479 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.751870 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.753388 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.759861 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.759871 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.759950 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.760178 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.760466 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.760570 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.760595 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-s4n5z" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.761373 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.770674 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.775890 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1fdac808-4983-49ff-9975-ebc0ad82bb27-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776106 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776226 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7379ac07-e6f6-4218-85c2-76883f50cf0e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776350 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7379ac07-e6f6-4218-85c2-76883f50cf0e-config-data\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776443 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776475 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1fdac808-4983-49ff-9975-ebc0ad82bb27-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776497 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b69qd\" (UniqueName: \"kubernetes.io/projected/7379ac07-e6f6-4218-85c2-76883f50cf0e-kube-api-access-b69qd\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776516 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776539 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776558 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7379ac07-e6f6-4218-85c2-76883f50cf0e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776580 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776631 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1fdac808-4983-49ff-9975-ebc0ad82bb27-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776646 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1fdac808-4983-49ff-9975-ebc0ad82bb27-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776693 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7379ac07-e6f6-4218-85c2-76883f50cf0e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776724 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776745 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1fdac808-4983-49ff-9975-ebc0ad82bb27-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776763 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7379ac07-e6f6-4218-85c2-76883f50cf0e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776833 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776881 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.776926 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.777000 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.777026 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5wm7\" (UniqueName: \"kubernetes.io/projected/1fdac808-4983-49ff-9975-ebc0ad82bb27-kube-api-access-x5wm7\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.878213 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7379ac07-e6f6-4218-85c2-76883f50cf0e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.878451 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.878475 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1fdac808-4983-49ff-9975-ebc0ad82bb27-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.878494 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7379ac07-e6f6-4218-85c2-76883f50cf0e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.878510 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.878975 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.878531 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879054 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879368 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879399 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879397 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1fdac808-4983-49ff-9975-ebc0ad82bb27-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879444 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5wm7\" (UniqueName: \"kubernetes.io/projected/1fdac808-4983-49ff-9975-ebc0ad82bb27-kube-api-access-x5wm7\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879527 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1fdac808-4983-49ff-9975-ebc0ad82bb27-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879576 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879625 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7379ac07-e6f6-4218-85c2-76883f50cf0e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879675 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7379ac07-e6f6-4218-85c2-76883f50cf0e-config-data\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879737 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879771 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1fdac808-4983-49ff-9975-ebc0ad82bb27-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879801 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b69qd\" (UniqueName: \"kubernetes.io/projected/7379ac07-e6f6-4218-85c2-76883f50cf0e-kube-api-access-b69qd\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879827 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879857 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879883 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7379ac07-e6f6-4218-85c2-76883f50cf0e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879886 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879913 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.879986 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1fdac808-4983-49ff-9975-ebc0ad82bb27-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.880003 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.880377 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7379ac07-e6f6-4218-85c2-76883f50cf0e-config-data\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.880011 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1fdac808-4983-49ff-9975-ebc0ad82bb27-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.880564 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.880773 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1fdac808-4983-49ff-9975-ebc0ad82bb27-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.880800 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7379ac07-e6f6-4218-85c2-76883f50cf0e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.881003 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1fdac808-4983-49ff-9975-ebc0ad82bb27-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.880007 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7379ac07-e6f6-4218-85c2-76883f50cf0e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.881130 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.882772 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1fdac808-4983-49ff-9975-ebc0ad82bb27-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.883288 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7379ac07-e6f6-4218-85c2-76883f50cf0e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.883795 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.883907 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7379ac07-e6f6-4218-85c2-76883f50cf0e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.886800 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.887749 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7379ac07-e6f6-4218-85c2-76883f50cf0e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.888405 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1fdac808-4983-49ff-9975-ebc0ad82bb27-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.894742 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1fdac808-4983-49ff-9975-ebc0ad82bb27-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.904413 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b69qd\" (UniqueName: \"kubernetes.io/projected/7379ac07-e6f6-4218-85c2-76883f50cf0e-kube-api-access-b69qd\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.912299 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5wm7\" (UniqueName: \"kubernetes.io/projected/1fdac808-4983-49ff-9975-ebc0ad82bb27-kube-api-access-x5wm7\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.927326 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1fdac808-4983-49ff-9975-ebc0ad82bb27\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:07 crc kubenswrapper[4969]: I1004 08:38:07.938054 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7379ac07-e6f6-4218-85c2-76883f50cf0e\") " pod="openstack/rabbitmq-server-0" Oct 04 08:38:08 crc kubenswrapper[4969]: I1004 08:38:08.074601 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:08 crc kubenswrapper[4969]: I1004 08:38:08.089697 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 08:38:08 crc kubenswrapper[4969]: E1004 08:38:08.310353 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod451dbcc5_7ad5_429a_8404_73b2c75466bd.slice/crio-485eb2d2641c2e39d7ce8e315ea71c55aeac2bad5e676213f98a5874110b835f\": RecentStats: unable to find data in memory cache]" Oct 04 08:38:08 crc kubenswrapper[4969]: I1004 08:38:08.549571 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 08:38:08 crc kubenswrapper[4969]: I1004 08:38:08.630480 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1fdac808-4983-49ff-9975-ebc0ad82bb27","Type":"ContainerStarted","Data":"705281501cf47bd6c5cf444c1487a9ee466e961694b309a30b93151a1be2376d"} Oct 04 08:38:09 crc kubenswrapper[4969]: I1004 08:38:09.075503 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1547b86a-ce65-4135-a8a5-957017c2271c" path="/var/lib/kubelet/pods/1547b86a-ce65-4135-a8a5-957017c2271c/volumes" Oct 04 08:38:19 crc kubenswrapper[4969]: I1004 08:38:19.666658 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:38:19 crc kubenswrapper[4969]: I1004 08:38:19.667174 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:38:22 crc kubenswrapper[4969]: W1004 08:38:22.783068 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7379ac07_e6f6_4218_85c2_76883f50cf0e.slice/crio-67f0c87c3ef37fdfb26a7f0a1bf404b6b3c62c01ccebc517070d2fa8fe684a69 WatchSource:0}: Error finding container 67f0c87c3ef37fdfb26a7f0a1bf404b6b3c62c01ccebc517070d2fa8fe684a69: Status 404 returned error can't find the container with id 67f0c87c3ef37fdfb26a7f0a1bf404b6b3c62c01ccebc517070d2fa8fe684a69 Oct 04 08:38:22 crc kubenswrapper[4969]: I1004 08:38:22.853330 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eaebec4-b71b-409d-b91e-1993605ce209" path="/var/lib/kubelet/pods/6eaebec4-b71b-409d-b91e-1993605ce209/volumes" Oct 04 08:38:22 crc kubenswrapper[4969]: E1004 08:38:22.854640 4969 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="13.8s" Oct 04 08:38:22 crc kubenswrapper[4969]: I1004 08:38:22.855103 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7379ac07-e6f6-4218-85c2-76883f50cf0e","Type":"ContainerStarted","Data":"67f0c87c3ef37fdfb26a7f0a1bf404b6b3c62c01ccebc517070d2fa8fe684a69"} Oct 04 08:38:22 crc kubenswrapper[4969]: I1004 08:38:22.887168 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 08:38:23 crc kubenswrapper[4969]: I1004 08:38:23.843394 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1fdac808-4983-49ff-9975-ebc0ad82bb27","Type":"ContainerStarted","Data":"07bf8c837d385d315f7092186bccf9272cd84b5005961c45ee4c0ebd5ba57514"} Oct 04 08:38:24 crc kubenswrapper[4969]: I1004 08:38:24.868927 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7379ac07-e6f6-4218-85c2-76883f50cf0e","Type":"ContainerStarted","Data":"c5214913d9eb360e83a3278939f6f82db43480fa4b52ec11c2200b8ad2a08fe8"} Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.736800 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84c9d6c57f-kq5lp"] Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.739220 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.742786 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.749364 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84c9d6c57f-kq5lp"] Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.810102 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-nb\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.810378 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh8wv\" (UniqueName: \"kubernetes.io/projected/836c6ebd-93c8-4238-9baa-d817651aa2c6-kube-api-access-sh8wv\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.810402 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-svc\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.810477 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-config\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.810553 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-swift-storage-0\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.810611 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-openstack-edpm-ipam\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.810668 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-sb\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.861473 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84c9d6c57f-kq5lp"] Oct 04 08:38:34 crc kubenswrapper[4969]: E1004 08:38:34.862191 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-sh8wv openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" podUID="836c6ebd-93c8-4238-9baa-d817651aa2c6" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.895757 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bd8b95f47-sjpqk"] Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.897858 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.911961 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-config\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.912027 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-swift-storage-0\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.912079 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-openstack-edpm-ipam\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.912136 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-sb\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.912178 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-nb\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.912194 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh8wv\" (UniqueName: \"kubernetes.io/projected/836c6ebd-93c8-4238-9baa-d817651aa2c6-kube-api-access-sh8wv\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.912216 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-svc\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.912921 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-config\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.913194 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-svc\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.913299 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-swift-storage-0\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.913936 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-nb\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.914440 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-sb\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.914583 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-openstack-edpm-ipam\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.917448 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd8b95f47-sjpqk"] Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.936731 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh8wv\" (UniqueName: \"kubernetes.io/projected/836c6ebd-93c8-4238-9baa-d817651aa2c6-kube-api-access-sh8wv\") pod \"dnsmasq-dns-84c9d6c57f-kq5lp\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:34 crc kubenswrapper[4969]: I1004 08:38:34.989860 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.004149 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.014066 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9g7g\" (UniqueName: \"kubernetes.io/projected/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-kube-api-access-z9g7g\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.014173 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-dns-svc\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.014235 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-dns-swift-storage-0\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.014275 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-config\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.014294 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-openstack-edpm-ipam\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.014339 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-ovsdbserver-nb\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.014361 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-ovsdbserver-sb\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.115407 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-config\") pod \"836c6ebd-93c8-4238-9baa-d817651aa2c6\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.115491 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh8wv\" (UniqueName: \"kubernetes.io/projected/836c6ebd-93c8-4238-9baa-d817651aa2c6-kube-api-access-sh8wv\") pod \"836c6ebd-93c8-4238-9baa-d817651aa2c6\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.115558 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-openstack-edpm-ipam\") pod \"836c6ebd-93c8-4238-9baa-d817651aa2c6\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.115617 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-nb\") pod \"836c6ebd-93c8-4238-9baa-d817651aa2c6\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.115713 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-swift-storage-0\") pod \"836c6ebd-93c8-4238-9baa-d817651aa2c6\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.115736 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-svc\") pod \"836c6ebd-93c8-4238-9baa-d817651aa2c6\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.115786 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-sb\") pod \"836c6ebd-93c8-4238-9baa-d817651aa2c6\" (UID: \"836c6ebd-93c8-4238-9baa-d817651aa2c6\") " Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.116347 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-dns-svc\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.116387 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-dns-swift-storage-0\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.116476 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-config\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.116497 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-openstack-edpm-ipam\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.116615 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-ovsdbserver-nb\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.116668 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-ovsdbserver-sb\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.116772 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9g7g\" (UniqueName: \"kubernetes.io/projected/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-kube-api-access-z9g7g\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.116792 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "836c6ebd-93c8-4238-9baa-d817651aa2c6" (UID: "836c6ebd-93c8-4238-9baa-d817651aa2c6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.117213 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.117405 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "836c6ebd-93c8-4238-9baa-d817651aa2c6" (UID: "836c6ebd-93c8-4238-9baa-d817651aa2c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.119085 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-ovsdbserver-sb\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.119729 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-ovsdbserver-nb\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.120808 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "836c6ebd-93c8-4238-9baa-d817651aa2c6" (UID: "836c6ebd-93c8-4238-9baa-d817651aa2c6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.121085 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "836c6ebd-93c8-4238-9baa-d817651aa2c6" (UID: "836c6ebd-93c8-4238-9baa-d817651aa2c6"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.121177 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-dns-svc\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.121389 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-config\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.123806 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "836c6ebd-93c8-4238-9baa-d817651aa2c6" (UID: "836c6ebd-93c8-4238-9baa-d817651aa2c6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.123831 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-config" (OuterVolumeSpecName: "config") pod "836c6ebd-93c8-4238-9baa-d817651aa2c6" (UID: "836c6ebd-93c8-4238-9baa-d817651aa2c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.124684 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-dns-swift-storage-0\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.128093 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-openstack-edpm-ipam\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.131637 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/836c6ebd-93c8-4238-9baa-d817651aa2c6-kube-api-access-sh8wv" (OuterVolumeSpecName: "kube-api-access-sh8wv") pod "836c6ebd-93c8-4238-9baa-d817651aa2c6" (UID: "836c6ebd-93c8-4238-9baa-d817651aa2c6"). InnerVolumeSpecName "kube-api-access-sh8wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.148526 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9g7g\" (UniqueName: \"kubernetes.io/projected/342bf40f-63d2-47cd-82c8-5c2aec79d3d2-kube-api-access-z9g7g\") pod \"dnsmasq-dns-bd8b95f47-sjpqk\" (UID: \"342bf40f-63d2-47cd-82c8-5c2aec79d3d2\") " pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.218463 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.219646 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.219683 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh8wv\" (UniqueName: \"kubernetes.io/projected/836c6ebd-93c8-4238-9baa-d817651aa2c6-kube-api-access-sh8wv\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.219697 4969 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.219707 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.219718 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.219728 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836c6ebd-93c8-4238-9baa-d817651aa2c6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.683174 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd8b95f47-sjpqk"] Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.999093 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" event={"ID":"342bf40f-63d2-47cd-82c8-5c2aec79d3d2","Type":"ContainerStarted","Data":"6de612f69337d5bc4c350e94eaf9ca20414379dd6f7e456c91b84a29012f9998"} Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.999121 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c9d6c57f-kq5lp" Oct 04 08:38:35 crc kubenswrapper[4969]: I1004 08:38:35.999143 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" event={"ID":"342bf40f-63d2-47cd-82c8-5c2aec79d3d2","Type":"ContainerStarted","Data":"379f1a0cdd880679d2e2c3260298239a44f594908d363badd4103bf87eaeb689"} Oct 04 08:38:36 crc kubenswrapper[4969]: I1004 08:38:36.096788 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84c9d6c57f-kq5lp"] Oct 04 08:38:36 crc kubenswrapper[4969]: I1004 08:38:36.109389 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84c9d6c57f-kq5lp"] Oct 04 08:38:37 crc kubenswrapper[4969]: I1004 08:38:37.009316 4969 generic.go:334] "Generic (PLEG): container finished" podID="342bf40f-63d2-47cd-82c8-5c2aec79d3d2" containerID="6de612f69337d5bc4c350e94eaf9ca20414379dd6f7e456c91b84a29012f9998" exitCode=0 Oct 04 08:38:37 crc kubenswrapper[4969]: I1004 08:38:37.009396 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" event={"ID":"342bf40f-63d2-47cd-82c8-5c2aec79d3d2","Type":"ContainerDied","Data":"6de612f69337d5bc4c350e94eaf9ca20414379dd6f7e456c91b84a29012f9998"} Oct 04 08:38:37 crc kubenswrapper[4969]: I1004 08:38:37.068812 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="836c6ebd-93c8-4238-9baa-d817651aa2c6" path="/var/lib/kubelet/pods/836c6ebd-93c8-4238-9baa-d817651aa2c6/volumes" Oct 04 08:38:38 crc kubenswrapper[4969]: I1004 08:38:38.020555 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" event={"ID":"342bf40f-63d2-47cd-82c8-5c2aec79d3d2","Type":"ContainerStarted","Data":"324697d99cb295558d174aa91253d8a932f4875f011fb2cb80948f67b3cf72c4"} Oct 04 08:38:38 crc kubenswrapper[4969]: I1004 08:38:38.020775 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:38 crc kubenswrapper[4969]: I1004 08:38:38.044147 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" podStartSLOduration=4.044131016 podStartE2EDuration="4.044131016s" podCreationTimestamp="2025-10-04 08:38:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:38:38.042052485 +0000 UTC m=+1345.796321299" watchObservedRunningTime="2025-10-04 08:38:38.044131016 +0000 UTC m=+1345.798399830" Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.221876 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bd8b95f47-sjpqk" Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.306933 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6df8b76f9c-knxjd"] Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.307167 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" podUID="ddb0ee51-df73-4216-ad98-2efcf1529efa" containerName="dnsmasq-dns" containerID="cri-o://50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab" gracePeriod=10 Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.838928 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.938507 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-swift-storage-0\") pod \"ddb0ee51-df73-4216-ad98-2efcf1529efa\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.938632 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-config\") pod \"ddb0ee51-df73-4216-ad98-2efcf1529efa\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.938673 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnx7b\" (UniqueName: \"kubernetes.io/projected/ddb0ee51-df73-4216-ad98-2efcf1529efa-kube-api-access-dnx7b\") pod \"ddb0ee51-df73-4216-ad98-2efcf1529efa\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.938712 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-sb\") pod \"ddb0ee51-df73-4216-ad98-2efcf1529efa\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.938751 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-nb\") pod \"ddb0ee51-df73-4216-ad98-2efcf1529efa\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.938822 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-svc\") pod \"ddb0ee51-df73-4216-ad98-2efcf1529efa\" (UID: \"ddb0ee51-df73-4216-ad98-2efcf1529efa\") " Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.949686 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddb0ee51-df73-4216-ad98-2efcf1529efa-kube-api-access-dnx7b" (OuterVolumeSpecName: "kube-api-access-dnx7b") pod "ddb0ee51-df73-4216-ad98-2efcf1529efa" (UID: "ddb0ee51-df73-4216-ad98-2efcf1529efa"). InnerVolumeSpecName "kube-api-access-dnx7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.992761 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ddb0ee51-df73-4216-ad98-2efcf1529efa" (UID: "ddb0ee51-df73-4216-ad98-2efcf1529efa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:45 crc kubenswrapper[4969]: I1004 08:38:45.997729 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-config" (OuterVolumeSpecName: "config") pod "ddb0ee51-df73-4216-ad98-2efcf1529efa" (UID: "ddb0ee51-df73-4216-ad98-2efcf1529efa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.015726 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ddb0ee51-df73-4216-ad98-2efcf1529efa" (UID: "ddb0ee51-df73-4216-ad98-2efcf1529efa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.017104 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ddb0ee51-df73-4216-ad98-2efcf1529efa" (UID: "ddb0ee51-df73-4216-ad98-2efcf1529efa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.026406 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ddb0ee51-df73-4216-ad98-2efcf1529efa" (UID: "ddb0ee51-df73-4216-ad98-2efcf1529efa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.041768 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.041802 4969 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.041812 4969 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.041823 4969 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.041834 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb0ee51-df73-4216-ad98-2efcf1529efa-config\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.041842 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnx7b\" (UniqueName: \"kubernetes.io/projected/ddb0ee51-df73-4216-ad98-2efcf1529efa-kube-api-access-dnx7b\") on node \"crc\" DevicePath \"\"" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.110326 4969 generic.go:334] "Generic (PLEG): container finished" podID="ddb0ee51-df73-4216-ad98-2efcf1529efa" containerID="50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab" exitCode=0 Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.110388 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" event={"ID":"ddb0ee51-df73-4216-ad98-2efcf1529efa","Type":"ContainerDied","Data":"50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab"} Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.110480 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" event={"ID":"ddb0ee51-df73-4216-ad98-2efcf1529efa","Type":"ContainerDied","Data":"f48e37fefe04c31ac1b9b82317a1dcc55cb74134ab09de3ec5303b80cd9c4d3e"} Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.110511 4969 scope.go:117] "RemoveContainer" containerID="50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.110915 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df8b76f9c-knxjd" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.132112 4969 scope.go:117] "RemoveContainer" containerID="b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.157241 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6df8b76f9c-knxjd"] Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.166289 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6df8b76f9c-knxjd"] Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.182920 4969 scope.go:117] "RemoveContainer" containerID="50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab" Oct 04 08:38:46 crc kubenswrapper[4969]: E1004 08:38:46.185462 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab\": container with ID starting with 50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab not found: ID does not exist" containerID="50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.185502 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab"} err="failed to get container status \"50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab\": rpc error: code = NotFound desc = could not find container \"50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab\": container with ID starting with 50ae2362b84960c14f8b3aa6235974637e59e4304909f2c135185a7b06cf11ab not found: ID does not exist" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.185528 4969 scope.go:117] "RemoveContainer" containerID="b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7" Oct 04 08:38:46 crc kubenswrapper[4969]: E1004 08:38:46.187858 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7\": container with ID starting with b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7 not found: ID does not exist" containerID="b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7" Oct 04 08:38:46 crc kubenswrapper[4969]: I1004 08:38:46.187889 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7"} err="failed to get container status \"b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7\": rpc error: code = NotFound desc = could not find container \"b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7\": container with ID starting with b04a5aab3671936d779c4e03ece25b78144871295fc6d00f03e088487069e7c7 not found: ID does not exist" Oct 04 08:38:47 crc kubenswrapper[4969]: I1004 08:38:47.072995 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddb0ee51-df73-4216-ad98-2efcf1529efa" path="/var/lib/kubelet/pods/ddb0ee51-df73-4216-ad98-2efcf1529efa/volumes" Oct 04 08:38:49 crc kubenswrapper[4969]: I1004 08:38:49.666159 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:38:49 crc kubenswrapper[4969]: I1004 08:38:49.666476 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:38:56 crc kubenswrapper[4969]: I1004 08:38:56.260290 4969 generic.go:334] "Generic (PLEG): container finished" podID="1fdac808-4983-49ff-9975-ebc0ad82bb27" containerID="07bf8c837d385d315f7092186bccf9272cd84b5005961c45ee4c0ebd5ba57514" exitCode=0 Oct 04 08:38:56 crc kubenswrapper[4969]: I1004 08:38:56.260394 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1fdac808-4983-49ff-9975-ebc0ad82bb27","Type":"ContainerDied","Data":"07bf8c837d385d315f7092186bccf9272cd84b5005961c45ee4c0ebd5ba57514"} Oct 04 08:38:57 crc kubenswrapper[4969]: I1004 08:38:57.273612 4969 generic.go:334] "Generic (PLEG): container finished" podID="7379ac07-e6f6-4218-85c2-76883f50cf0e" containerID="c5214913d9eb360e83a3278939f6f82db43480fa4b52ec11c2200b8ad2a08fe8" exitCode=0 Oct 04 08:38:57 crc kubenswrapper[4969]: I1004 08:38:57.273709 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7379ac07-e6f6-4218-85c2-76883f50cf0e","Type":"ContainerDied","Data":"c5214913d9eb360e83a3278939f6f82db43480fa4b52ec11c2200b8ad2a08fe8"} Oct 04 08:38:57 crc kubenswrapper[4969]: I1004 08:38:57.277324 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1fdac808-4983-49ff-9975-ebc0ad82bb27","Type":"ContainerStarted","Data":"de8cd2d8b7daa12349fff897d2e1905767af5caa44bda9a3d23592deb20b9bdb"} Oct 04 08:38:57 crc kubenswrapper[4969]: I1004 08:38:57.278094 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:38:57 crc kubenswrapper[4969]: I1004 08:38:57.324492 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.324475135 podStartE2EDuration="50.324475135s" podCreationTimestamp="2025-10-04 08:38:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:38:57.323366477 +0000 UTC m=+1365.077635291" watchObservedRunningTime="2025-10-04 08:38:57.324475135 +0000 UTC m=+1365.078743949" Oct 04 08:38:58 crc kubenswrapper[4969]: I1004 08:38:58.290701 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7379ac07-e6f6-4218-85c2-76883f50cf0e","Type":"ContainerStarted","Data":"618b3773c47bdec0e1f3461f4e870ddb6b960b4f8549ea5f34ee307e883108f5"} Oct 04 08:38:58 crc kubenswrapper[4969]: I1004 08:38:58.291388 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 08:38:58 crc kubenswrapper[4969]: I1004 08:38:58.328083 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=51.328052038 podStartE2EDuration="51.328052038s" podCreationTimestamp="2025-10-04 08:38:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:38:58.320198326 +0000 UTC m=+1366.074467180" watchObservedRunningTime="2025-10-04 08:38:58.328052038 +0000 UTC m=+1366.082320902" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.388453 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z"] Oct 04 08:38:59 crc kubenswrapper[4969]: E1004 08:38:59.389091 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb0ee51-df73-4216-ad98-2efcf1529efa" containerName="dnsmasq-dns" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.389103 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb0ee51-df73-4216-ad98-2efcf1529efa" containerName="dnsmasq-dns" Oct 04 08:38:59 crc kubenswrapper[4969]: E1004 08:38:59.389116 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb0ee51-df73-4216-ad98-2efcf1529efa" containerName="init" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.389121 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb0ee51-df73-4216-ad98-2efcf1529efa" containerName="init" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.389298 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddb0ee51-df73-4216-ad98-2efcf1529efa" containerName="dnsmasq-dns" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.389987 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.394727 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.395110 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.423323 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.423739 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.466605 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z"] Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.574689 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.574738 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.575083 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.575256 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rcfh\" (UniqueName: \"kubernetes.io/projected/a4b37224-1961-43f0-bbc9-efbc635bc063-kube-api-access-9rcfh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.677405 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.677482 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.678097 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.678771 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rcfh\" (UniqueName: \"kubernetes.io/projected/a4b37224-1961-43f0-bbc9-efbc635bc063-kube-api-access-9rcfh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.684012 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.684088 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.684728 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.698332 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rcfh\" (UniqueName: \"kubernetes.io/projected/a4b37224-1961-43f0-bbc9-efbc635bc063-kube-api-access-9rcfh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:38:59 crc kubenswrapper[4969]: I1004 08:38:59.760058 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:39:00 crc kubenswrapper[4969]: I1004 08:39:00.371812 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z"] Oct 04 08:39:00 crc kubenswrapper[4969]: W1004 08:39:00.375090 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4b37224_1961_43f0_bbc9_efbc635bc063.slice/crio-b05d611185ada35681c6c318844628c63beb0f591c52fb3948b6f18a9d654b9c WatchSource:0}: Error finding container b05d611185ada35681c6c318844628c63beb0f591c52fb3948b6f18a9d654b9c: Status 404 returned error can't find the container with id b05d611185ada35681c6c318844628c63beb0f591c52fb3948b6f18a9d654b9c Oct 04 08:39:00 crc kubenswrapper[4969]: I1004 08:39:00.377669 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 08:39:01 crc kubenswrapper[4969]: I1004 08:39:01.326772 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" event={"ID":"a4b37224-1961-43f0-bbc9-efbc635bc063","Type":"ContainerStarted","Data":"b05d611185ada35681c6c318844628c63beb0f591c52fb3948b6f18a9d654b9c"} Oct 04 08:39:08 crc kubenswrapper[4969]: I1004 08:39:08.076456 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="1fdac808-4983-49ff-9975-ebc0ad82bb27" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.226:5671: connect: connection refused" Oct 04 08:39:08 crc kubenswrapper[4969]: I1004 08:39:08.092955 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7379ac07-e6f6-4218-85c2-76883f50cf0e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.227:5671: connect: connection refused" Oct 04 08:39:13 crc kubenswrapper[4969]: I1004 08:39:13.472210 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" event={"ID":"a4b37224-1961-43f0-bbc9-efbc635bc063","Type":"ContainerStarted","Data":"06bd950cbc9aff76c820a94aeef98f4c314d9a4d9c420adeb7d238ac1edcd8a1"} Oct 04 08:39:18 crc kubenswrapper[4969]: I1004 08:39:18.077699 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 08:39:18 crc kubenswrapper[4969]: I1004 08:39:18.093751 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 08:39:18 crc kubenswrapper[4969]: I1004 08:39:18.116327 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" podStartSLOduration=6.9483669500000005 podStartE2EDuration="19.11629343s" podCreationTimestamp="2025-10-04 08:38:59 +0000 UTC" firstStartedPulling="2025-10-04 08:39:00.377488973 +0000 UTC m=+1368.131757787" lastFinishedPulling="2025-10-04 08:39:12.545415443 +0000 UTC m=+1380.299684267" observedRunningTime="2025-10-04 08:39:13.494752222 +0000 UTC m=+1381.249021046" watchObservedRunningTime="2025-10-04 08:39:18.11629343 +0000 UTC m=+1385.870562284" Oct 04 08:39:19 crc kubenswrapper[4969]: I1004 08:39:19.667059 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:39:19 crc kubenswrapper[4969]: I1004 08:39:19.667468 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:39:19 crc kubenswrapper[4969]: I1004 08:39:19.667534 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:39:19 crc kubenswrapper[4969]: I1004 08:39:19.668407 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"139722c0c58580b4c2a5dc3b9a83a0c499c007e51329c9719d5fb5495e97f91c"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:39:19 crc kubenswrapper[4969]: I1004 08:39:19.668504 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://139722c0c58580b4c2a5dc3b9a83a0c499c007e51329c9719d5fb5495e97f91c" gracePeriod=600 Oct 04 08:39:20 crc kubenswrapper[4969]: I1004 08:39:20.560796 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="139722c0c58580b4c2a5dc3b9a83a0c499c007e51329c9719d5fb5495e97f91c" exitCode=0 Oct 04 08:39:20 crc kubenswrapper[4969]: I1004 08:39:20.560892 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"139722c0c58580b4c2a5dc3b9a83a0c499c007e51329c9719d5fb5495e97f91c"} Oct 04 08:39:20 crc kubenswrapper[4969]: I1004 08:39:20.561312 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6"} Oct 04 08:39:20 crc kubenswrapper[4969]: I1004 08:39:20.561335 4969 scope.go:117] "RemoveContainer" containerID="f952a2138e627189e2388732fb9b24c40ce1c3b15ba69934279b103aa755d5b4" Oct 04 08:39:27 crc kubenswrapper[4969]: I1004 08:39:27.150177 4969 scope.go:117] "RemoveContainer" containerID="7293ab8d08343f9c68c67fb2c9a6ca5eae2fbf92ee2915c9b9c3d3620b5a04f8" Oct 04 08:39:29 crc kubenswrapper[4969]: I1004 08:39:29.679513 4969 generic.go:334] "Generic (PLEG): container finished" podID="a4b37224-1961-43f0-bbc9-efbc635bc063" containerID="06bd950cbc9aff76c820a94aeef98f4c314d9a4d9c420adeb7d238ac1edcd8a1" exitCode=0 Oct 04 08:39:29 crc kubenswrapper[4969]: I1004 08:39:29.679647 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" event={"ID":"a4b37224-1961-43f0-bbc9-efbc635bc063","Type":"ContainerDied","Data":"06bd950cbc9aff76c820a94aeef98f4c314d9a4d9c420adeb7d238ac1edcd8a1"} Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.222957 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.336986 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-repo-setup-combined-ca-bundle\") pod \"a4b37224-1961-43f0-bbc9-efbc635bc063\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.337109 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-ssh-key\") pod \"a4b37224-1961-43f0-bbc9-efbc635bc063\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.337321 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rcfh\" (UniqueName: \"kubernetes.io/projected/a4b37224-1961-43f0-bbc9-efbc635bc063-kube-api-access-9rcfh\") pod \"a4b37224-1961-43f0-bbc9-efbc635bc063\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.337364 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-inventory\") pod \"a4b37224-1961-43f0-bbc9-efbc635bc063\" (UID: \"a4b37224-1961-43f0-bbc9-efbc635bc063\") " Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.342964 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a4b37224-1961-43f0-bbc9-efbc635bc063" (UID: "a4b37224-1961-43f0-bbc9-efbc635bc063"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.344847 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4b37224-1961-43f0-bbc9-efbc635bc063-kube-api-access-9rcfh" (OuterVolumeSpecName: "kube-api-access-9rcfh") pod "a4b37224-1961-43f0-bbc9-efbc635bc063" (UID: "a4b37224-1961-43f0-bbc9-efbc635bc063"). InnerVolumeSpecName "kube-api-access-9rcfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.368602 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-inventory" (OuterVolumeSpecName: "inventory") pod "a4b37224-1961-43f0-bbc9-efbc635bc063" (UID: "a4b37224-1961-43f0-bbc9-efbc635bc063"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.386223 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a4b37224-1961-43f0-bbc9-efbc635bc063" (UID: "a4b37224-1961-43f0-bbc9-efbc635bc063"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.440500 4969 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.440535 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.440547 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rcfh\" (UniqueName: \"kubernetes.io/projected/a4b37224-1961-43f0-bbc9-efbc635bc063-kube-api-access-9rcfh\") on node \"crc\" DevicePath \"\"" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.440561 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4b37224-1961-43f0-bbc9-efbc635bc063-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.704532 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" event={"ID":"a4b37224-1961-43f0-bbc9-efbc635bc063","Type":"ContainerDied","Data":"b05d611185ada35681c6c318844628c63beb0f591c52fb3948b6f18a9d654b9c"} Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.704575 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b05d611185ada35681c6c318844628c63beb0f591c52fb3948b6f18a9d654b9c" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.704632 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.840167 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg"] Oct 04 08:39:31 crc kubenswrapper[4969]: E1004 08:39:31.841090 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4b37224-1961-43f0-bbc9-efbc635bc063" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.841142 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4b37224-1961-43f0-bbc9-efbc635bc063" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.841686 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4b37224-1961-43f0-bbc9-efbc635bc063" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.843135 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.846084 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.846271 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.859725 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg"] Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.859986 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.862088 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.951186 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j9fsg\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.951285 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j9fsg\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:31 crc kubenswrapper[4969]: I1004 08:39:31.951322 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl7k6\" (UniqueName: \"kubernetes.io/projected/562cc378-72b8-42ea-9822-4a5aa6cebbb6-kube-api-access-nl7k6\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j9fsg\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:32 crc kubenswrapper[4969]: I1004 08:39:32.053855 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j9fsg\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:32 crc kubenswrapper[4969]: I1004 08:39:32.054013 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j9fsg\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:32 crc kubenswrapper[4969]: I1004 08:39:32.054960 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl7k6\" (UniqueName: \"kubernetes.io/projected/562cc378-72b8-42ea-9822-4a5aa6cebbb6-kube-api-access-nl7k6\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j9fsg\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:32 crc kubenswrapper[4969]: I1004 08:39:32.059534 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j9fsg\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:32 crc kubenswrapper[4969]: I1004 08:39:32.070323 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j9fsg\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:32 crc kubenswrapper[4969]: I1004 08:39:32.076482 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl7k6\" (UniqueName: \"kubernetes.io/projected/562cc378-72b8-42ea-9822-4a5aa6cebbb6-kube-api-access-nl7k6\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j9fsg\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:32 crc kubenswrapper[4969]: I1004 08:39:32.171907 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:32 crc kubenswrapper[4969]: I1004 08:39:32.762485 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg"] Oct 04 08:39:33 crc kubenswrapper[4969]: I1004 08:39:33.731729 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" event={"ID":"562cc378-72b8-42ea-9822-4a5aa6cebbb6","Type":"ContainerStarted","Data":"b8abaacb304078352018720b109a06db704cf8339c3eb4a922b8ee88d058a0e9"} Oct 04 08:39:35 crc kubenswrapper[4969]: I1004 08:39:35.756267 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" event={"ID":"562cc378-72b8-42ea-9822-4a5aa6cebbb6","Type":"ContainerStarted","Data":"17ec8195d33118cf02be6e40a4350183b647a24bee333d0de7c8c5ecdc59f173"} Oct 04 08:39:35 crc kubenswrapper[4969]: I1004 08:39:35.787565 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" podStartSLOduration=2.883410118 podStartE2EDuration="4.787547091s" podCreationTimestamp="2025-10-04 08:39:31 +0000 UTC" firstStartedPulling="2025-10-04 08:39:32.768468039 +0000 UTC m=+1400.522736853" lastFinishedPulling="2025-10-04 08:39:34.672604972 +0000 UTC m=+1402.426873826" observedRunningTime="2025-10-04 08:39:35.778652514 +0000 UTC m=+1403.532921318" watchObservedRunningTime="2025-10-04 08:39:35.787547091 +0000 UTC m=+1403.541815895" Oct 04 08:39:37 crc kubenswrapper[4969]: I1004 08:39:37.779926 4969 generic.go:334] "Generic (PLEG): container finished" podID="562cc378-72b8-42ea-9822-4a5aa6cebbb6" containerID="17ec8195d33118cf02be6e40a4350183b647a24bee333d0de7c8c5ecdc59f173" exitCode=0 Oct 04 08:39:37 crc kubenswrapper[4969]: I1004 08:39:37.780020 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" event={"ID":"562cc378-72b8-42ea-9822-4a5aa6cebbb6","Type":"ContainerDied","Data":"17ec8195d33118cf02be6e40a4350183b647a24bee333d0de7c8c5ecdc59f173"} Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.277021 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.348662 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-ssh-key\") pod \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.348730 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-inventory\") pod \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.348958 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl7k6\" (UniqueName: \"kubernetes.io/projected/562cc378-72b8-42ea-9822-4a5aa6cebbb6-kube-api-access-nl7k6\") pod \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\" (UID: \"562cc378-72b8-42ea-9822-4a5aa6cebbb6\") " Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.357652 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/562cc378-72b8-42ea-9822-4a5aa6cebbb6-kube-api-access-nl7k6" (OuterVolumeSpecName: "kube-api-access-nl7k6") pod "562cc378-72b8-42ea-9822-4a5aa6cebbb6" (UID: "562cc378-72b8-42ea-9822-4a5aa6cebbb6"). InnerVolumeSpecName "kube-api-access-nl7k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.389254 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-inventory" (OuterVolumeSpecName: "inventory") pod "562cc378-72b8-42ea-9822-4a5aa6cebbb6" (UID: "562cc378-72b8-42ea-9822-4a5aa6cebbb6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.393179 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "562cc378-72b8-42ea-9822-4a5aa6cebbb6" (UID: "562cc378-72b8-42ea-9822-4a5aa6cebbb6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.451544 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.451789 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562cc378-72b8-42ea-9822-4a5aa6cebbb6-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.451799 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl7k6\" (UniqueName: \"kubernetes.io/projected/562cc378-72b8-42ea-9822-4a5aa6cebbb6-kube-api-access-nl7k6\") on node \"crc\" DevicePath \"\"" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.805484 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" event={"ID":"562cc378-72b8-42ea-9822-4a5aa6cebbb6","Type":"ContainerDied","Data":"b8abaacb304078352018720b109a06db704cf8339c3eb4a922b8ee88d058a0e9"} Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.805521 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8abaacb304078352018720b109a06db704cf8339c3eb4a922b8ee88d058a0e9" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.805583 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j9fsg" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.904582 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw"] Oct 04 08:39:39 crc kubenswrapper[4969]: E1004 08:39:39.905145 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562cc378-72b8-42ea-9822-4a5aa6cebbb6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.905164 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="562cc378-72b8-42ea-9822-4a5aa6cebbb6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.905446 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="562cc378-72b8-42ea-9822-4a5aa6cebbb6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.906257 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.909644 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.909641 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.909822 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.910314 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.913692 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw"] Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.961864 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.961990 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.962071 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:39 crc kubenswrapper[4969]: I1004 08:39:39.962175 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld6k4\" (UniqueName: \"kubernetes.io/projected/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-kube-api-access-ld6k4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.064291 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld6k4\" (UniqueName: \"kubernetes.io/projected/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-kube-api-access-ld6k4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.064525 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.064645 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.064738 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.069833 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.071201 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.072283 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.093901 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld6k4\" (UniqueName: \"kubernetes.io/projected/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-kube-api-access-ld6k4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.224293 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.784886 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw"] Oct 04 08:39:40 crc kubenswrapper[4969]: W1004 08:39:40.790186 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e5dcf0e_60ad_4ddf_b0a7_e504f4566ce7.slice/crio-400b8a572e8c3fdab613450b28cc21594766279d194c61e89968acf1fe2dbc20 WatchSource:0}: Error finding container 400b8a572e8c3fdab613450b28cc21594766279d194c61e89968acf1fe2dbc20: Status 404 returned error can't find the container with id 400b8a572e8c3fdab613450b28cc21594766279d194c61e89968acf1fe2dbc20 Oct 04 08:39:40 crc kubenswrapper[4969]: I1004 08:39:40.816307 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" event={"ID":"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7","Type":"ContainerStarted","Data":"400b8a572e8c3fdab613450b28cc21594766279d194c61e89968acf1fe2dbc20"} Oct 04 08:39:41 crc kubenswrapper[4969]: I1004 08:39:41.833329 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" event={"ID":"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7","Type":"ContainerStarted","Data":"799215b9bad6967faeb8958eedc3c5b881f4325c5955cceab1f06bd3413f3a8f"} Oct 04 08:39:41 crc kubenswrapper[4969]: I1004 08:39:41.852323 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" podStartSLOduration=2.410349994 podStartE2EDuration="2.852304522s" podCreationTimestamp="2025-10-04 08:39:39 +0000 UTC" firstStartedPulling="2025-10-04 08:39:40.792777925 +0000 UTC m=+1408.547046749" lastFinishedPulling="2025-10-04 08:39:41.234732423 +0000 UTC m=+1408.989001277" observedRunningTime="2025-10-04 08:39:41.847497226 +0000 UTC m=+1409.601766140" watchObservedRunningTime="2025-10-04 08:39:41.852304522 +0000 UTC m=+1409.606573336" Oct 04 08:40:17 crc kubenswrapper[4969]: I1004 08:40:17.801937 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7jfb4"] Oct 04 08:40:17 crc kubenswrapper[4969]: I1004 08:40:17.805721 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:17 crc kubenswrapper[4969]: I1004 08:40:17.829933 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jfb4"] Oct 04 08:40:17 crc kubenswrapper[4969]: I1004 08:40:17.964408 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-catalog-content\") pod \"redhat-operators-7jfb4\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:17 crc kubenswrapper[4969]: I1004 08:40:17.964622 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq9kc\" (UniqueName: \"kubernetes.io/projected/b04cc5a4-ee66-4623-aa49-91d2f759bc91-kube-api-access-nq9kc\") pod \"redhat-operators-7jfb4\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:17 crc kubenswrapper[4969]: I1004 08:40:17.964734 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-utilities\") pod \"redhat-operators-7jfb4\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:18 crc kubenswrapper[4969]: I1004 08:40:18.066734 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq9kc\" (UniqueName: \"kubernetes.io/projected/b04cc5a4-ee66-4623-aa49-91d2f759bc91-kube-api-access-nq9kc\") pod \"redhat-operators-7jfb4\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:18 crc kubenswrapper[4969]: I1004 08:40:18.066878 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-utilities\") pod \"redhat-operators-7jfb4\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:18 crc kubenswrapper[4969]: I1004 08:40:18.067064 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-catalog-content\") pod \"redhat-operators-7jfb4\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:18 crc kubenswrapper[4969]: I1004 08:40:18.067876 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-catalog-content\") pod \"redhat-operators-7jfb4\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:18 crc kubenswrapper[4969]: I1004 08:40:18.068056 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-utilities\") pod \"redhat-operators-7jfb4\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:18 crc kubenswrapper[4969]: I1004 08:40:18.098292 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq9kc\" (UniqueName: \"kubernetes.io/projected/b04cc5a4-ee66-4623-aa49-91d2f759bc91-kube-api-access-nq9kc\") pod \"redhat-operators-7jfb4\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:18 crc kubenswrapper[4969]: I1004 08:40:18.131624 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:18 crc kubenswrapper[4969]: I1004 08:40:18.611664 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jfb4"] Oct 04 08:40:19 crc kubenswrapper[4969]: I1004 08:40:19.331965 4969 generic.go:334] "Generic (PLEG): container finished" podID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerID="f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0" exitCode=0 Oct 04 08:40:19 crc kubenswrapper[4969]: I1004 08:40:19.332307 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jfb4" event={"ID":"b04cc5a4-ee66-4623-aa49-91d2f759bc91","Type":"ContainerDied","Data":"f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0"} Oct 04 08:40:19 crc kubenswrapper[4969]: I1004 08:40:19.332363 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jfb4" event={"ID":"b04cc5a4-ee66-4623-aa49-91d2f759bc91","Type":"ContainerStarted","Data":"26dee076db27f08594c1472e0adf25ccf6686bbbb67f837b60226d4e88c25f22"} Oct 04 08:40:20 crc kubenswrapper[4969]: I1004 08:40:20.347018 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jfb4" event={"ID":"b04cc5a4-ee66-4623-aa49-91d2f759bc91","Type":"ContainerStarted","Data":"a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5"} Oct 04 08:40:22 crc kubenswrapper[4969]: I1004 08:40:22.381048 4969 generic.go:334] "Generic (PLEG): container finished" podID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerID="a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5" exitCode=0 Oct 04 08:40:22 crc kubenswrapper[4969]: I1004 08:40:22.381112 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jfb4" event={"ID":"b04cc5a4-ee66-4623-aa49-91d2f759bc91","Type":"ContainerDied","Data":"a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5"} Oct 04 08:40:24 crc kubenswrapper[4969]: I1004 08:40:24.418543 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jfb4" event={"ID":"b04cc5a4-ee66-4623-aa49-91d2f759bc91","Type":"ContainerStarted","Data":"f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396"} Oct 04 08:40:24 crc kubenswrapper[4969]: I1004 08:40:24.454579 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7jfb4" podStartSLOduration=3.658168796 podStartE2EDuration="7.454552453s" podCreationTimestamp="2025-10-04 08:40:17 +0000 UTC" firstStartedPulling="2025-10-04 08:40:19.336900617 +0000 UTC m=+1447.091169441" lastFinishedPulling="2025-10-04 08:40:23.133284284 +0000 UTC m=+1450.887553098" observedRunningTime="2025-10-04 08:40:24.438684586 +0000 UTC m=+1452.192953440" watchObservedRunningTime="2025-10-04 08:40:24.454552453 +0000 UTC m=+1452.208821307" Oct 04 08:40:27 crc kubenswrapper[4969]: I1004 08:40:27.233122 4969 scope.go:117] "RemoveContainer" containerID="b7e88c823b7a23fb328c159dd65d4e4806e654ea40682085f11dd7ccb3d94b19" Oct 04 08:40:27 crc kubenswrapper[4969]: I1004 08:40:27.304853 4969 scope.go:117] "RemoveContainer" containerID="85771b674ebd673fcaeda8fa43cea00728395310592d2f7c048d0767b1c51055" Oct 04 08:40:27 crc kubenswrapper[4969]: I1004 08:40:27.335055 4969 scope.go:117] "RemoveContainer" containerID="001e2338fea808e3665ffd16d9def28352878cb322c3ca5065371fdfd6537ced" Oct 04 08:40:28 crc kubenswrapper[4969]: I1004 08:40:28.132515 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:28 crc kubenswrapper[4969]: I1004 08:40:28.132575 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:29 crc kubenswrapper[4969]: I1004 08:40:29.191956 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7jfb4" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="registry-server" probeResult="failure" output=< Oct 04 08:40:29 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 08:40:29 crc kubenswrapper[4969]: > Oct 04 08:40:39 crc kubenswrapper[4969]: I1004 08:40:39.186847 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7jfb4" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="registry-server" probeResult="failure" output=< Oct 04 08:40:39 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 08:40:39 crc kubenswrapper[4969]: > Oct 04 08:40:48 crc kubenswrapper[4969]: I1004 08:40:48.196293 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:48 crc kubenswrapper[4969]: I1004 08:40:48.256864 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:49 crc kubenswrapper[4969]: I1004 08:40:49.015591 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7jfb4"] Oct 04 08:40:49 crc kubenswrapper[4969]: I1004 08:40:49.723779 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7jfb4" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="registry-server" containerID="cri-o://f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396" gracePeriod=2 Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.226046 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.364743 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-utilities\") pod \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.364851 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq9kc\" (UniqueName: \"kubernetes.io/projected/b04cc5a4-ee66-4623-aa49-91d2f759bc91-kube-api-access-nq9kc\") pod \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.364890 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-catalog-content\") pod \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\" (UID: \"b04cc5a4-ee66-4623-aa49-91d2f759bc91\") " Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.365728 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-utilities" (OuterVolumeSpecName: "utilities") pod "b04cc5a4-ee66-4623-aa49-91d2f759bc91" (UID: "b04cc5a4-ee66-4623-aa49-91d2f759bc91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.375815 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b04cc5a4-ee66-4623-aa49-91d2f759bc91-kube-api-access-nq9kc" (OuterVolumeSpecName: "kube-api-access-nq9kc") pod "b04cc5a4-ee66-4623-aa49-91d2f759bc91" (UID: "b04cc5a4-ee66-4623-aa49-91d2f759bc91"). InnerVolumeSpecName "kube-api-access-nq9kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.466838 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.466872 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq9kc\" (UniqueName: \"kubernetes.io/projected/b04cc5a4-ee66-4623-aa49-91d2f759bc91-kube-api-access-nq9kc\") on node \"crc\" DevicePath \"\"" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.470493 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b04cc5a4-ee66-4623-aa49-91d2f759bc91" (UID: "b04cc5a4-ee66-4623-aa49-91d2f759bc91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.568678 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04cc5a4-ee66-4623-aa49-91d2f759bc91-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.735372 4969 generic.go:334] "Generic (PLEG): container finished" podID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerID="f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396" exitCode=0 Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.735459 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jfb4" event={"ID":"b04cc5a4-ee66-4623-aa49-91d2f759bc91","Type":"ContainerDied","Data":"f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396"} Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.735493 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jfb4" event={"ID":"b04cc5a4-ee66-4623-aa49-91d2f759bc91","Type":"ContainerDied","Data":"26dee076db27f08594c1472e0adf25ccf6686bbbb67f837b60226d4e88c25f22"} Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.735540 4969 scope.go:117] "RemoveContainer" containerID="f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.735743 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jfb4" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.760863 4969 scope.go:117] "RemoveContainer" containerID="a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.787786 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7jfb4"] Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.796644 4969 scope.go:117] "RemoveContainer" containerID="f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.800058 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7jfb4"] Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.850613 4969 scope.go:117] "RemoveContainer" containerID="f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396" Oct 04 08:40:50 crc kubenswrapper[4969]: E1004 08:40:50.850942 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396\": container with ID starting with f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396 not found: ID does not exist" containerID="f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.850968 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396"} err="failed to get container status \"f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396\": rpc error: code = NotFound desc = could not find container \"f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396\": container with ID starting with f780b0c1b7cd0116e1aa17bcf42c3329630abc158705c0fdb90892f0238f4396 not found: ID does not exist" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.850988 4969 scope.go:117] "RemoveContainer" containerID="a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5" Oct 04 08:40:50 crc kubenswrapper[4969]: E1004 08:40:50.851381 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5\": container with ID starting with a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5 not found: ID does not exist" containerID="a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.851518 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5"} err="failed to get container status \"a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5\": rpc error: code = NotFound desc = could not find container \"a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5\": container with ID starting with a7b6ca75ec0f72a1074b73d9bffa336148d302fab25f82f440bc5ba172778ea5 not found: ID does not exist" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.851612 4969 scope.go:117] "RemoveContainer" containerID="f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0" Oct 04 08:40:50 crc kubenswrapper[4969]: E1004 08:40:50.851981 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0\": container with ID starting with f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0 not found: ID does not exist" containerID="f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0" Oct 04 08:40:50 crc kubenswrapper[4969]: I1004 08:40:50.852002 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0"} err="failed to get container status \"f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0\": rpc error: code = NotFound desc = could not find container \"f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0\": container with ID starting with f008d637b135d376b97b3daafecf9ee30f3bc84db1e70f0ce3e57d65e42e2de0 not found: ID does not exist" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.073754 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" path="/var/lib/kubelet/pods/b04cc5a4-ee66-4623-aa49-91d2f759bc91/volumes" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.422203 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-phcmj"] Oct 04 08:40:51 crc kubenswrapper[4969]: E1004 08:40:51.422757 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="registry-server" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.422785 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="registry-server" Oct 04 08:40:51 crc kubenswrapper[4969]: E1004 08:40:51.422821 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="extract-content" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.422829 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="extract-content" Oct 04 08:40:51 crc kubenswrapper[4969]: E1004 08:40:51.422853 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="extract-utilities" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.422866 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="extract-utilities" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.423215 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="b04cc5a4-ee66-4623-aa49-91d2f759bc91" containerName="registry-server" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.425773 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.439888 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-phcmj"] Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.591371 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-utilities\") pod \"redhat-marketplace-phcmj\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.591818 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-catalog-content\") pod \"redhat-marketplace-phcmj\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.591879 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxt9s\" (UniqueName: \"kubernetes.io/projected/40624840-b0ca-4e7e-bb29-5e9a868363cf-kube-api-access-pxt9s\") pod \"redhat-marketplace-phcmj\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.694620 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-utilities\") pod \"redhat-marketplace-phcmj\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.694735 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-catalog-content\") pod \"redhat-marketplace-phcmj\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.694755 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxt9s\" (UniqueName: \"kubernetes.io/projected/40624840-b0ca-4e7e-bb29-5e9a868363cf-kube-api-access-pxt9s\") pod \"redhat-marketplace-phcmj\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.695542 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-utilities\") pod \"redhat-marketplace-phcmj\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.696049 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-catalog-content\") pod \"redhat-marketplace-phcmj\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.717753 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxt9s\" (UniqueName: \"kubernetes.io/projected/40624840-b0ca-4e7e-bb29-5e9a868363cf-kube-api-access-pxt9s\") pod \"redhat-marketplace-phcmj\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:51 crc kubenswrapper[4969]: I1004 08:40:51.762544 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:40:52 crc kubenswrapper[4969]: I1004 08:40:52.283248 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-phcmj"] Oct 04 08:40:52 crc kubenswrapper[4969]: W1004 08:40:52.294851 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40624840_b0ca_4e7e_bb29_5e9a868363cf.slice/crio-5d3967f99eb52d90ad9f4ad082b79605a40f756f58051cc3151a001ed65836f4 WatchSource:0}: Error finding container 5d3967f99eb52d90ad9f4ad082b79605a40f756f58051cc3151a001ed65836f4: Status 404 returned error can't find the container with id 5d3967f99eb52d90ad9f4ad082b79605a40f756f58051cc3151a001ed65836f4 Oct 04 08:40:52 crc kubenswrapper[4969]: I1004 08:40:52.765508 4969 generic.go:334] "Generic (PLEG): container finished" podID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerID="377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd" exitCode=0 Oct 04 08:40:52 crc kubenswrapper[4969]: I1004 08:40:52.765585 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phcmj" event={"ID":"40624840-b0ca-4e7e-bb29-5e9a868363cf","Type":"ContainerDied","Data":"377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd"} Oct 04 08:40:52 crc kubenswrapper[4969]: I1004 08:40:52.765631 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phcmj" event={"ID":"40624840-b0ca-4e7e-bb29-5e9a868363cf","Type":"ContainerStarted","Data":"5d3967f99eb52d90ad9f4ad082b79605a40f756f58051cc3151a001ed65836f4"} Oct 04 08:40:54 crc kubenswrapper[4969]: I1004 08:40:54.790516 4969 generic.go:334] "Generic (PLEG): container finished" podID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerID="45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30" exitCode=0 Oct 04 08:40:54 crc kubenswrapper[4969]: I1004 08:40:54.790606 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phcmj" event={"ID":"40624840-b0ca-4e7e-bb29-5e9a868363cf","Type":"ContainerDied","Data":"45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30"} Oct 04 08:40:55 crc kubenswrapper[4969]: I1004 08:40:55.805168 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phcmj" event={"ID":"40624840-b0ca-4e7e-bb29-5e9a868363cf","Type":"ContainerStarted","Data":"307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72"} Oct 04 08:40:55 crc kubenswrapper[4969]: I1004 08:40:55.837616 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-phcmj" podStartSLOduration=2.372048777 podStartE2EDuration="4.83758948s" podCreationTimestamp="2025-10-04 08:40:51 +0000 UTC" firstStartedPulling="2025-10-04 08:40:52.768296054 +0000 UTC m=+1480.522564908" lastFinishedPulling="2025-10-04 08:40:55.233836797 +0000 UTC m=+1482.988105611" observedRunningTime="2025-10-04 08:40:55.825738291 +0000 UTC m=+1483.580007125" watchObservedRunningTime="2025-10-04 08:40:55.83758948 +0000 UTC m=+1483.591858344" Oct 04 08:41:01 crc kubenswrapper[4969]: I1004 08:41:01.763291 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:41:01 crc kubenswrapper[4969]: I1004 08:41:01.765106 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:41:01 crc kubenswrapper[4969]: I1004 08:41:01.819929 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:41:01 crc kubenswrapper[4969]: I1004 08:41:01.943566 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:41:02 crc kubenswrapper[4969]: I1004 08:41:02.070382 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-phcmj"] Oct 04 08:41:03 crc kubenswrapper[4969]: I1004 08:41:03.909578 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-phcmj" podUID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerName="registry-server" containerID="cri-o://307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72" gracePeriod=2 Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.389496 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.507881 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-catalog-content\") pod \"40624840-b0ca-4e7e-bb29-5e9a868363cf\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.508119 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-utilities\") pod \"40624840-b0ca-4e7e-bb29-5e9a868363cf\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.508222 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxt9s\" (UniqueName: \"kubernetes.io/projected/40624840-b0ca-4e7e-bb29-5e9a868363cf-kube-api-access-pxt9s\") pod \"40624840-b0ca-4e7e-bb29-5e9a868363cf\" (UID: \"40624840-b0ca-4e7e-bb29-5e9a868363cf\") " Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.509198 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-utilities" (OuterVolumeSpecName: "utilities") pod "40624840-b0ca-4e7e-bb29-5e9a868363cf" (UID: "40624840-b0ca-4e7e-bb29-5e9a868363cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.515226 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40624840-b0ca-4e7e-bb29-5e9a868363cf-kube-api-access-pxt9s" (OuterVolumeSpecName: "kube-api-access-pxt9s") pod "40624840-b0ca-4e7e-bb29-5e9a868363cf" (UID: "40624840-b0ca-4e7e-bb29-5e9a868363cf"). InnerVolumeSpecName "kube-api-access-pxt9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.535533 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40624840-b0ca-4e7e-bb29-5e9a868363cf" (UID: "40624840-b0ca-4e7e-bb29-5e9a868363cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.610810 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.610872 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxt9s\" (UniqueName: \"kubernetes.io/projected/40624840-b0ca-4e7e-bb29-5e9a868363cf-kube-api-access-pxt9s\") on node \"crc\" DevicePath \"\"" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.610891 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40624840-b0ca-4e7e-bb29-5e9a868363cf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.923774 4969 generic.go:334] "Generic (PLEG): container finished" podID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerID="307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72" exitCode=0 Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.923847 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phcmj" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.923870 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phcmj" event={"ID":"40624840-b0ca-4e7e-bb29-5e9a868363cf","Type":"ContainerDied","Data":"307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72"} Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.924495 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phcmj" event={"ID":"40624840-b0ca-4e7e-bb29-5e9a868363cf","Type":"ContainerDied","Data":"5d3967f99eb52d90ad9f4ad082b79605a40f756f58051cc3151a001ed65836f4"} Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.924503 4969 scope.go:117] "RemoveContainer" containerID="307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.972915 4969 scope.go:117] "RemoveContainer" containerID="45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30" Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.977260 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-phcmj"] Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.995946 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-phcmj"] Oct 04 08:41:04 crc kubenswrapper[4969]: I1004 08:41:04.997843 4969 scope.go:117] "RemoveContainer" containerID="377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd" Oct 04 08:41:05 crc kubenswrapper[4969]: I1004 08:41:05.070181 4969 scope.go:117] "RemoveContainer" containerID="307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72" Oct 04 08:41:05 crc kubenswrapper[4969]: E1004 08:41:05.070754 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72\": container with ID starting with 307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72 not found: ID does not exist" containerID="307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72" Oct 04 08:41:05 crc kubenswrapper[4969]: I1004 08:41:05.070795 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72"} err="failed to get container status \"307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72\": rpc error: code = NotFound desc = could not find container \"307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72\": container with ID starting with 307116940c0fc61a051324624e70e0515a995699becb0c08aa0292c1dbc32e72 not found: ID does not exist" Oct 04 08:41:05 crc kubenswrapper[4969]: I1004 08:41:05.070819 4969 scope.go:117] "RemoveContainer" containerID="45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30" Oct 04 08:41:05 crc kubenswrapper[4969]: E1004 08:41:05.071405 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30\": container with ID starting with 45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30 not found: ID does not exist" containerID="45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30" Oct 04 08:41:05 crc kubenswrapper[4969]: I1004 08:41:05.071458 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30"} err="failed to get container status \"45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30\": rpc error: code = NotFound desc = could not find container \"45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30\": container with ID starting with 45cbf8f22c2d80ba1a2a3f38bf7efad0479cbf722500b592859c6253759edf30 not found: ID does not exist" Oct 04 08:41:05 crc kubenswrapper[4969]: I1004 08:41:05.071477 4969 scope.go:117] "RemoveContainer" containerID="377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd" Oct 04 08:41:05 crc kubenswrapper[4969]: E1004 08:41:05.071978 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd\": container with ID starting with 377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd not found: ID does not exist" containerID="377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd" Oct 04 08:41:05 crc kubenswrapper[4969]: I1004 08:41:05.072013 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd"} err="failed to get container status \"377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd\": rpc error: code = NotFound desc = could not find container \"377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd\": container with ID starting with 377ae099a7d344dce32645206f19a4d8b471781360168c222894aae74b85afbd not found: ID does not exist" Oct 04 08:41:05 crc kubenswrapper[4969]: I1004 08:41:05.080239 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40624840-b0ca-4e7e-bb29-5e9a868363cf" path="/var/lib/kubelet/pods/40624840-b0ca-4e7e-bb29-5e9a868363cf/volumes" Oct 04 08:41:49 crc kubenswrapper[4969]: I1004 08:41:49.666679 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:41:49 crc kubenswrapper[4969]: I1004 08:41:49.667219 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:42:19 crc kubenswrapper[4969]: I1004 08:42:19.667199 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:42:19 crc kubenswrapper[4969]: I1004 08:42:19.667738 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:42:49 crc kubenswrapper[4969]: I1004 08:42:49.667177 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:42:49 crc kubenswrapper[4969]: I1004 08:42:49.667842 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:42:49 crc kubenswrapper[4969]: I1004 08:42:49.667905 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:42:49 crc kubenswrapper[4969]: I1004 08:42:49.669051 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:42:49 crc kubenswrapper[4969]: I1004 08:42:49.669177 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" gracePeriod=600 Oct 04 08:42:49 crc kubenswrapper[4969]: E1004 08:42:49.799777 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:42:50 crc kubenswrapper[4969]: I1004 08:42:50.162713 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" exitCode=0 Oct 04 08:42:50 crc kubenswrapper[4969]: I1004 08:42:50.162772 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6"} Oct 04 08:42:50 crc kubenswrapper[4969]: I1004 08:42:50.162817 4969 scope.go:117] "RemoveContainer" containerID="139722c0c58580b4c2a5dc3b9a83a0c499c007e51329c9719d5fb5495e97f91c" Oct 04 08:42:50 crc kubenswrapper[4969]: I1004 08:42:50.163625 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:42:50 crc kubenswrapper[4969]: E1004 08:42:50.164035 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:43:04 crc kubenswrapper[4969]: I1004 08:43:04.055272 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:43:04 crc kubenswrapper[4969]: E1004 08:43:04.056326 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:43:14 crc kubenswrapper[4969]: I1004 08:43:14.459168 4969 generic.go:334] "Generic (PLEG): container finished" podID="6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7" containerID="799215b9bad6967faeb8958eedc3c5b881f4325c5955cceab1f06bd3413f3a8f" exitCode=0 Oct 04 08:43:14 crc kubenswrapper[4969]: I1004 08:43:14.459263 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" event={"ID":"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7","Type":"ContainerDied","Data":"799215b9bad6967faeb8958eedc3c5b881f4325c5955cceab1f06bd3413f3a8f"} Oct 04 08:43:15 crc kubenswrapper[4969]: I1004 08:43:15.057008 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:43:15 crc kubenswrapper[4969]: E1004 08:43:15.058023 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:43:15 crc kubenswrapper[4969]: I1004 08:43:15.944457 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.032641 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-bootstrap-combined-ca-bundle\") pod \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.032697 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-ssh-key\") pod \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.032730 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-inventory\") pod \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.032854 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld6k4\" (UniqueName: \"kubernetes.io/projected/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-kube-api-access-ld6k4\") pod \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\" (UID: \"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7\") " Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.052146 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-xlvgm"] Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.053456 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-kube-api-access-ld6k4" (OuterVolumeSpecName: "kube-api-access-ld6k4") pod "6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7" (UID: "6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7"). InnerVolumeSpecName "kube-api-access-ld6k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.053887 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7" (UID: "6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.061441 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-xlvgm"] Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.071127 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7" (UID: "6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.072592 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-inventory" (OuterVolumeSpecName: "inventory") pod "6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7" (UID: "6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.138204 4969 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.138237 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.138245 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.138257 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld6k4\" (UniqueName: \"kubernetes.io/projected/6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7-kube-api-access-ld6k4\") on node \"crc\" DevicePath \"\"" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.485060 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" event={"ID":"6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7","Type":"ContainerDied","Data":"400b8a572e8c3fdab613450b28cc21594766279d194c61e89968acf1fe2dbc20"} Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.485509 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="400b8a572e8c3fdab613450b28cc21594766279d194c61e89968acf1fe2dbc20" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.485124 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.590245 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz"] Oct 04 08:43:16 crc kubenswrapper[4969]: E1004 08:43:16.590788 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerName="registry-server" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.590813 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerName="registry-server" Oct 04 08:43:16 crc kubenswrapper[4969]: E1004 08:43:16.590866 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.590876 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 08:43:16 crc kubenswrapper[4969]: E1004 08:43:16.590900 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerName="extract-utilities" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.590909 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerName="extract-utilities" Oct 04 08:43:16 crc kubenswrapper[4969]: E1004 08:43:16.590929 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerName="extract-content" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.590937 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerName="extract-content" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.591165 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.591190 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="40624840-b0ca-4e7e-bb29-5e9a868363cf" containerName="registry-server" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.592028 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.594097 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.594522 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.594712 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.594764 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.613907 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz"] Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.649630 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cxn7\" (UniqueName: \"kubernetes.io/projected/a71c86d5-acae-401d-bea7-8cee0956d990-kube-api-access-6cxn7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.649684 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.649826 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.752035 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.752159 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cxn7\" (UniqueName: \"kubernetes.io/projected/a71c86d5-acae-401d-bea7-8cee0956d990-kube-api-access-6cxn7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.752193 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.760210 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.765869 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.767878 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cxn7\" (UniqueName: \"kubernetes.io/projected/a71c86d5-acae-401d-bea7-8cee0956d990-kube-api-access-6cxn7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:16 crc kubenswrapper[4969]: I1004 08:43:16.914390 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:43:17 crc kubenswrapper[4969]: I1004 08:43:17.073840 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b39fae4-bbfe-466b-a08a-52119cf4e4a6" path="/var/lib/kubelet/pods/9b39fae4-bbfe-466b-a08a-52119cf4e4a6/volumes" Oct 04 08:43:17 crc kubenswrapper[4969]: I1004 08:43:17.449083 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz"] Oct 04 08:43:17 crc kubenswrapper[4969]: I1004 08:43:17.495875 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" event={"ID":"a71c86d5-acae-401d-bea7-8cee0956d990","Type":"ContainerStarted","Data":"06dd2c0b14682d57a9d4c3f2399b2b8728a168b1231ab47351c72b21120c51db"} Oct 04 08:43:18 crc kubenswrapper[4969]: I1004 08:43:18.508962 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" event={"ID":"a71c86d5-acae-401d-bea7-8cee0956d990","Type":"ContainerStarted","Data":"61b9a8bf81ac42170cf14d2710e2e38c98a9324506c10cdced9e72a98a532ae8"} Oct 04 08:43:18 crc kubenswrapper[4969]: I1004 08:43:18.532033 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" podStartSLOduration=1.8947894330000001 podStartE2EDuration="2.532014458s" podCreationTimestamp="2025-10-04 08:43:16 +0000 UTC" firstStartedPulling="2025-10-04 08:43:17.451023548 +0000 UTC m=+1625.205292372" lastFinishedPulling="2025-10-04 08:43:18.088248573 +0000 UTC m=+1625.842517397" observedRunningTime="2025-10-04 08:43:18.527043747 +0000 UTC m=+1626.281312591" watchObservedRunningTime="2025-10-04 08:43:18.532014458 +0000 UTC m=+1626.286283272" Oct 04 08:43:22 crc kubenswrapper[4969]: I1004 08:43:22.035745 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-7qcfn"] Oct 04 08:43:22 crc kubenswrapper[4969]: I1004 08:43:22.048742 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-7qcfn"] Oct 04 08:43:23 crc kubenswrapper[4969]: I1004 08:43:23.031299 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-5zcjl"] Oct 04 08:43:23 crc kubenswrapper[4969]: I1004 08:43:23.040299 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-cg6hf"] Oct 04 08:43:23 crc kubenswrapper[4969]: I1004 08:43:23.049453 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-5zcjl"] Oct 04 08:43:23 crc kubenswrapper[4969]: I1004 08:43:23.068690 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f9ae792-20d2-4617-9dd4-3b8ad58b12db" path="/var/lib/kubelet/pods/2f9ae792-20d2-4617-9dd4-3b8ad58b12db/volumes" Oct 04 08:43:23 crc kubenswrapper[4969]: I1004 08:43:23.069963 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d23e05af-c71b-4c8d-ac33-0a0aafbf3e32" path="/var/lib/kubelet/pods/d23e05af-c71b-4c8d-ac33-0a0aafbf3e32/volumes" Oct 04 08:43:23 crc kubenswrapper[4969]: I1004 08:43:23.070538 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-cg6hf"] Oct 04 08:43:25 crc kubenswrapper[4969]: I1004 08:43:25.026947 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-5c31-account-create-cqhsm"] Oct 04 08:43:25 crc kubenswrapper[4969]: I1004 08:43:25.037890 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-5c31-account-create-cqhsm"] Oct 04 08:43:25 crc kubenswrapper[4969]: I1004 08:43:25.069799 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42d6b82d-a2fa-463e-8075-04bf086f145b" path="/var/lib/kubelet/pods/42d6b82d-a2fa-463e-8075-04bf086f145b/volumes" Oct 04 08:43:25 crc kubenswrapper[4969]: I1004 08:43:25.070518 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="472bf752-8d20-4b60-82f8-71c73e4f8843" path="/var/lib/kubelet/pods/472bf752-8d20-4b60-82f8-71c73e4f8843/volumes" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.595968 4969 scope.go:117] "RemoveContainer" containerID="c8da89e1e2530e1a8b0c314ecb957c582482a1301969f958bac691bad4b3aa85" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.626846 4969 scope.go:117] "RemoveContainer" containerID="b10977563312793ba1ddd5c043d43a0e09269b8825d400f40a0dca08696f9d45" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.696583 4969 scope.go:117] "RemoveContainer" containerID="f94b523751c4b5bba921ee496cbdac61cd6117da890ca466756822e6ced2f46c" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.726888 4969 scope.go:117] "RemoveContainer" containerID="3436b85a700dab71576f2b84084f24b062c0d248728d5ca6871242ea7986d689" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.765504 4969 scope.go:117] "RemoveContainer" containerID="13ae23401025f1c7e27c81052bff279d623ad948115ac8ca0af2652325a8d006" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.781990 4969 scope.go:117] "RemoveContainer" containerID="8a7f7c9681d88161014fb95f54a27fed0435322438935efcda4191c686af1022" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.801578 4969 scope.go:117] "RemoveContainer" containerID="739922c6c43eef8820ea06f48bc977eb83476eeeeca4745b06309ef31bfbda92" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.824593 4969 scope.go:117] "RemoveContainer" containerID="30e093e2106f123922932ff85d99396ce44570943dee67da2c8233ecb3fad031" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.868193 4969 scope.go:117] "RemoveContainer" containerID="5a4f7d989a81b1e07ece8912930b2c83ea8cab66f8d80948f68b6950842f204e" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.917505 4969 scope.go:117] "RemoveContainer" containerID="7fc94259ee054c59317135db33a8f1abdbf881dd61b963611392daed882ea91b" Oct 04 08:43:27 crc kubenswrapper[4969]: I1004 08:43:27.954318 4969 scope.go:117] "RemoveContainer" containerID="3902102d442378626146542a829f8ccfc51b5df600b8bb5eb0ba844cf03fb8b1" Oct 04 08:43:28 crc kubenswrapper[4969]: I1004 08:43:28.055239 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:43:28 crc kubenswrapper[4969]: E1004 08:43:28.055530 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:43:30 crc kubenswrapper[4969]: I1004 08:43:30.041899 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-0476-account-create-vjvkm"] Oct 04 08:43:30 crc kubenswrapper[4969]: I1004 08:43:30.055727 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-0476-account-create-vjvkm"] Oct 04 08:43:31 crc kubenswrapper[4969]: I1004 08:43:31.071983 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f213477-11ae-42a6-b4e7-376453c4f64a" path="/var/lib/kubelet/pods/8f213477-11ae-42a6-b4e7-376453c4f64a/volumes" Oct 04 08:43:34 crc kubenswrapper[4969]: I1004 08:43:34.053825 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-34b9-account-create-z7dgt"] Oct 04 08:43:34 crc kubenswrapper[4969]: I1004 08:43:34.069273 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a8d7-account-create-xgxvr"] Oct 04 08:43:34 crc kubenswrapper[4969]: I1004 08:43:34.081200 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-34b9-account-create-z7dgt"] Oct 04 08:43:34 crc kubenswrapper[4969]: I1004 08:43:34.091348 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a8d7-account-create-xgxvr"] Oct 04 08:43:35 crc kubenswrapper[4969]: I1004 08:43:35.070299 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3293da46-00e5-4183-a2ed-7c702651f3cc" path="/var/lib/kubelet/pods/3293da46-00e5-4183-a2ed-7c702651f3cc/volumes" Oct 04 08:43:35 crc kubenswrapper[4969]: I1004 08:43:35.071701 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8187bdb9-d868-4c45-bb7e-df0c38f70e9b" path="/var/lib/kubelet/pods/8187bdb9-d868-4c45-bb7e-df0c38f70e9b/volumes" Oct 04 08:43:41 crc kubenswrapper[4969]: I1004 08:43:41.055716 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:43:41 crc kubenswrapper[4969]: E1004 08:43:41.056532 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:43:54 crc kubenswrapper[4969]: I1004 08:43:54.037286 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-5rtdq"] Oct 04 08:43:54 crc kubenswrapper[4969]: I1004 08:43:54.051147 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-9ch7d"] Oct 04 08:43:54 crc kubenswrapper[4969]: I1004 08:43:54.078451 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:43:54 crc kubenswrapper[4969]: I1004 08:43:54.078872 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-5rtdq"] Oct 04 08:43:54 crc kubenswrapper[4969]: E1004 08:43:54.079989 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:43:54 crc kubenswrapper[4969]: I1004 08:43:54.090946 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-9ch7d"] Oct 04 08:43:54 crc kubenswrapper[4969]: I1004 08:43:54.100801 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-5bss9"] Oct 04 08:43:54 crc kubenswrapper[4969]: I1004 08:43:54.111842 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-5bss9"] Oct 04 08:43:55 crc kubenswrapper[4969]: I1004 08:43:55.070516 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36354e6c-6917-4239-9ad3-c2cb67c8f765" path="/var/lib/kubelet/pods/36354e6c-6917-4239-9ad3-c2cb67c8f765/volumes" Oct 04 08:43:55 crc kubenswrapper[4969]: I1004 08:43:55.072805 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49667965-1201-4bd7-8b23-adf9f35d866f" path="/var/lib/kubelet/pods/49667965-1201-4bd7-8b23-adf9f35d866f/volumes" Oct 04 08:43:55 crc kubenswrapper[4969]: I1004 08:43:55.073533 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="550752d6-efab-448d-8776-6f6031942c11" path="/var/lib/kubelet/pods/550752d6-efab-448d-8776-6f6031942c11/volumes" Oct 04 08:44:00 crc kubenswrapper[4969]: I1004 08:44:00.040917 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-2t48v"] Oct 04 08:44:00 crc kubenswrapper[4969]: I1004 08:44:00.058348 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-2t48v"] Oct 04 08:44:00 crc kubenswrapper[4969]: I1004 08:44:00.068209 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-7nbv5"] Oct 04 08:44:00 crc kubenswrapper[4969]: I1004 08:44:00.077004 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-7nbv5"] Oct 04 08:44:01 crc kubenswrapper[4969]: I1004 08:44:01.069627 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd901023-7f58-4004-bf0e-7df1600efbeb" path="/var/lib/kubelet/pods/bd901023-7f58-4004-bf0e-7df1600efbeb/volumes" Oct 04 08:44:01 crc kubenswrapper[4969]: I1004 08:44:01.072509 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed6a88af-0fc2-45f8-b584-451c87b14ae6" path="/var/lib/kubelet/pods/ed6a88af-0fc2-45f8-b584-451c87b14ae6/volumes" Oct 04 08:44:06 crc kubenswrapper[4969]: I1004 08:44:06.056089 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:44:06 crc kubenswrapper[4969]: E1004 08:44:06.056996 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:44:18 crc kubenswrapper[4969]: I1004 08:44:18.056043 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:44:18 crc kubenswrapper[4969]: E1004 08:44:18.057026 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:44:23 crc kubenswrapper[4969]: I1004 08:44:23.053947 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-rskkf"] Oct 04 08:44:23 crc kubenswrapper[4969]: I1004 08:44:23.080690 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-rskkf"] Oct 04 08:44:25 crc kubenswrapper[4969]: I1004 08:44:25.085779 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35c12eab-f598-4088-b689-ad192cd83887" path="/var/lib/kubelet/pods/35c12eab-f598-4088-b689-ad192cd83887/volumes" Oct 04 08:44:26 crc kubenswrapper[4969]: I1004 08:44:26.033265 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-0dbb-account-create-k45jh"] Oct 04 08:44:26 crc kubenswrapper[4969]: I1004 08:44:26.046287 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d9f0-account-create-5zlxv"] Oct 04 08:44:26 crc kubenswrapper[4969]: I1004 08:44:26.060687 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-0dbb-account-create-k45jh"] Oct 04 08:44:26 crc kubenswrapper[4969]: I1004 08:44:26.069876 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d9f0-account-create-5zlxv"] Oct 04 08:44:27 crc kubenswrapper[4969]: I1004 08:44:27.032684 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6ab2-account-create-mpxpw"] Oct 04 08:44:27 crc kubenswrapper[4969]: I1004 08:44:27.043956 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6ab2-account-create-mpxpw"] Oct 04 08:44:27 crc kubenswrapper[4969]: I1004 08:44:27.070948 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93ccd81b-783b-4c03-b564-425601b5df4c" path="/var/lib/kubelet/pods/93ccd81b-783b-4c03-b564-425601b5df4c/volumes" Oct 04 08:44:27 crc kubenswrapper[4969]: I1004 08:44:27.072474 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a873063d-e103-48c3-b85e-1b8b2e0ba249" path="/var/lib/kubelet/pods/a873063d-e103-48c3-b85e-1b8b2e0ba249/volumes" Oct 04 08:44:27 crc kubenswrapper[4969]: I1004 08:44:27.073148 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef5c63b3-e8f7-4b9a-bb7c-642516049ee5" path="/var/lib/kubelet/pods/ef5c63b3-e8f7-4b9a-bb7c-642516049ee5/volumes" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.120966 4969 scope.go:117] "RemoveContainer" containerID="6f07054d682b8c5f0df8bcce9ec08f7801b172020e6186e12ad4b4d2321ccc3c" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.146499 4969 scope.go:117] "RemoveContainer" containerID="72d3eea0e6a149ea0314050fc3f297c8f1a2f61001947c5edd930688eaf47004" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.205402 4969 scope.go:117] "RemoveContainer" containerID="f2a430039a366420a178b21f2dd8fe7f8320fadfcaf782a05c8b1dc2c1d5993e" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.302776 4969 scope.go:117] "RemoveContainer" containerID="a5f9b29e226f1f785d18020b72e1529e218c6e168a1c5a730852f368d851c3c4" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.371146 4969 scope.go:117] "RemoveContainer" containerID="9644a9f250a026f90c923338b9d3be04047805319fc90514b03999744717a5e5" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.400786 4969 scope.go:117] "RemoveContainer" containerID="3e6ffec15040977a3b6cc2dfc19a9cce09ee68281ab8c8e130437cc576655f49" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.491688 4969 scope.go:117] "RemoveContainer" containerID="bb8c5d8a95ca5c483cdb480b53f11cf6594af1c1a88be3da46bd52fc9832ea5f" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.512586 4969 scope.go:117] "RemoveContainer" containerID="a7a695ad30ea674847a4264ecfb2baeae12a9c427cd584bc0a8ccc33d1b4d8e8" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.545595 4969 scope.go:117] "RemoveContainer" containerID="b1f95ee6222b1871dc4a4a676b443bf3d0733ab09387a6662fd2ea3d4a585122" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.581791 4969 scope.go:117] "RemoveContainer" containerID="906fbe48d89804808f909ec418ecb7797887c8aea9472b1601387ae440795c67" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.616083 4969 scope.go:117] "RemoveContainer" containerID="b81188eb7053c4f9e2356cb249f9e3eefc37024c49e128545314631891193999" Oct 04 08:44:28 crc kubenswrapper[4969]: I1004 08:44:28.670168 4969 scope.go:117] "RemoveContainer" containerID="8558501ad2698336951a93011571b0c299db2d03fb0cad4750355cfce6089045" Oct 04 08:44:32 crc kubenswrapper[4969]: I1004 08:44:32.055579 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:44:32 crc kubenswrapper[4969]: E1004 08:44:32.056798 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:44:35 crc kubenswrapper[4969]: I1004 08:44:35.086815 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6bdc9"] Oct 04 08:44:35 crc kubenswrapper[4969]: I1004 08:44:35.093216 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6bdc9"] Oct 04 08:44:37 crc kubenswrapper[4969]: I1004 08:44:37.074475 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e20b7533-fc8d-41c4-be2d-d117a56fa332" path="/var/lib/kubelet/pods/e20b7533-fc8d-41c4-be2d-d117a56fa332/volumes" Oct 04 08:44:47 crc kubenswrapper[4969]: I1004 08:44:47.056096 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:44:47 crc kubenswrapper[4969]: E1004 08:44:47.057509 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:44:50 crc kubenswrapper[4969]: I1004 08:44:50.047762 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nx8dt"] Oct 04 08:44:50 crc kubenswrapper[4969]: I1004 08:44:50.064036 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nx8dt"] Oct 04 08:44:51 crc kubenswrapper[4969]: I1004 08:44:51.071743 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a60a257-4654-4bfd-b878-94733004ac3d" path="/var/lib/kubelet/pods/4a60a257-4654-4bfd-b878-94733004ac3d/volumes" Oct 04 08:44:59 crc kubenswrapper[4969]: I1004 08:44:59.055316 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:44:59 crc kubenswrapper[4969]: E1004 08:44:59.056311 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.173659 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8"] Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.176512 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.186139 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.186370 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.204154 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8"] Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.364908 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-config-volume\") pod \"collect-profiles-29326125-58fk8\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.365075 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v2kw\" (UniqueName: \"kubernetes.io/projected/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-kube-api-access-7v2kw\") pod \"collect-profiles-29326125-58fk8\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.365135 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-secret-volume\") pod \"collect-profiles-29326125-58fk8\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.466278 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-secret-volume\") pod \"collect-profiles-29326125-58fk8\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.466331 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-config-volume\") pod \"collect-profiles-29326125-58fk8\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.466471 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v2kw\" (UniqueName: \"kubernetes.io/projected/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-kube-api-access-7v2kw\") pod \"collect-profiles-29326125-58fk8\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.468205 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-config-volume\") pod \"collect-profiles-29326125-58fk8\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.476924 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-secret-volume\") pod \"collect-profiles-29326125-58fk8\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.482622 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v2kw\" (UniqueName: \"kubernetes.io/projected/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-kube-api-access-7v2kw\") pod \"collect-profiles-29326125-58fk8\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:00 crc kubenswrapper[4969]: I1004 08:45:00.531815 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:01 crc kubenswrapper[4969]: I1004 08:45:01.031213 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8"] Oct 04 08:45:01 crc kubenswrapper[4969]: I1004 08:45:01.754763 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" event={"ID":"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa","Type":"ContainerStarted","Data":"f61d3630bbda83374faa3d0d7b73f1eedd62c00d83d055900270d976c485026d"} Oct 04 08:45:01 crc kubenswrapper[4969]: I1004 08:45:01.755058 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" event={"ID":"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa","Type":"ContainerStarted","Data":"933b8a2e6792602be03694a90203dcfe1e1175358bc7c24c5d32de8319bc7e9c"} Oct 04 08:45:01 crc kubenswrapper[4969]: I1004 08:45:01.783246 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" podStartSLOduration=1.783225447 podStartE2EDuration="1.783225447s" podCreationTimestamp="2025-10-04 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 08:45:01.779366294 +0000 UTC m=+1729.533635128" watchObservedRunningTime="2025-10-04 08:45:01.783225447 +0000 UTC m=+1729.537494271" Oct 04 08:45:02 crc kubenswrapper[4969]: I1004 08:45:02.766496 4969 generic.go:334] "Generic (PLEG): container finished" podID="69cdc4bb-3b8e-4094-9ee6-fab85d869ffa" containerID="f61d3630bbda83374faa3d0d7b73f1eedd62c00d83d055900270d976c485026d" exitCode=0 Oct 04 08:45:02 crc kubenswrapper[4969]: I1004 08:45:02.766606 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" event={"ID":"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa","Type":"ContainerDied","Data":"f61d3630bbda83374faa3d0d7b73f1eedd62c00d83d055900270d976c485026d"} Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.166503 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.248455 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-config-volume\") pod \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.248556 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v2kw\" (UniqueName: \"kubernetes.io/projected/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-kube-api-access-7v2kw\") pod \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.248636 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-secret-volume\") pod \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\" (UID: \"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa\") " Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.250072 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-config-volume" (OuterVolumeSpecName: "config-volume") pod "69cdc4bb-3b8e-4094-9ee6-fab85d869ffa" (UID: "69cdc4bb-3b8e-4094-9ee6-fab85d869ffa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.254470 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "69cdc4bb-3b8e-4094-9ee6-fab85d869ffa" (UID: "69cdc4bb-3b8e-4094-9ee6-fab85d869ffa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.256843 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-kube-api-access-7v2kw" (OuterVolumeSpecName: "kube-api-access-7v2kw") pod "69cdc4bb-3b8e-4094-9ee6-fab85d869ffa" (UID: "69cdc4bb-3b8e-4094-9ee6-fab85d869ffa"). InnerVolumeSpecName "kube-api-access-7v2kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.349975 4969 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.350013 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v2kw\" (UniqueName: \"kubernetes.io/projected/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-kube-api-access-7v2kw\") on node \"crc\" DevicePath \"\"" Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.350026 4969 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.790880 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" event={"ID":"69cdc4bb-3b8e-4094-9ee6-fab85d869ffa","Type":"ContainerDied","Data":"933b8a2e6792602be03694a90203dcfe1e1175358bc7c24c5d32de8319bc7e9c"} Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.791212 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="933b8a2e6792602be03694a90203dcfe1e1175358bc7c24c5d32de8319bc7e9c" Oct 04 08:45:04 crc kubenswrapper[4969]: I1004 08:45:04.791375 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8" Oct 04 08:45:08 crc kubenswrapper[4969]: I1004 08:45:08.066683 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ktnkt"] Oct 04 08:45:08 crc kubenswrapper[4969]: I1004 08:45:08.082982 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ktnkt"] Oct 04 08:45:09 crc kubenswrapper[4969]: I1004 08:45:09.079617 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1bc4bef-f750-4b65-96c5-af742379cf87" path="/var/lib/kubelet/pods/b1bc4bef-f750-4b65-96c5-af742379cf87/volumes" Oct 04 08:45:10 crc kubenswrapper[4969]: I1004 08:45:10.056066 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:45:10 crc kubenswrapper[4969]: E1004 08:45:10.056444 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:45:15 crc kubenswrapper[4969]: I1004 08:45:15.926981 4969 generic.go:334] "Generic (PLEG): container finished" podID="a71c86d5-acae-401d-bea7-8cee0956d990" containerID="61b9a8bf81ac42170cf14d2710e2e38c98a9324506c10cdced9e72a98a532ae8" exitCode=0 Oct 04 08:45:15 crc kubenswrapper[4969]: I1004 08:45:15.927060 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" event={"ID":"a71c86d5-acae-401d-bea7-8cee0956d990","Type":"ContainerDied","Data":"61b9a8bf81ac42170cf14d2710e2e38c98a9324506c10cdced9e72a98a532ae8"} Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.377293 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.538503 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cxn7\" (UniqueName: \"kubernetes.io/projected/a71c86d5-acae-401d-bea7-8cee0956d990-kube-api-access-6cxn7\") pod \"a71c86d5-acae-401d-bea7-8cee0956d990\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.538707 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-ssh-key\") pod \"a71c86d5-acae-401d-bea7-8cee0956d990\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.538806 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-inventory\") pod \"a71c86d5-acae-401d-bea7-8cee0956d990\" (UID: \"a71c86d5-acae-401d-bea7-8cee0956d990\") " Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.546502 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a71c86d5-acae-401d-bea7-8cee0956d990-kube-api-access-6cxn7" (OuterVolumeSpecName: "kube-api-access-6cxn7") pod "a71c86d5-acae-401d-bea7-8cee0956d990" (UID: "a71c86d5-acae-401d-bea7-8cee0956d990"). InnerVolumeSpecName "kube-api-access-6cxn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.568246 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-inventory" (OuterVolumeSpecName: "inventory") pod "a71c86d5-acae-401d-bea7-8cee0956d990" (UID: "a71c86d5-acae-401d-bea7-8cee0956d990"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.587100 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a71c86d5-acae-401d-bea7-8cee0956d990" (UID: "a71c86d5-acae-401d-bea7-8cee0956d990"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.641347 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.641377 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71c86d5-acae-401d-bea7-8cee0956d990-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.641387 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cxn7\" (UniqueName: \"kubernetes.io/projected/a71c86d5-acae-401d-bea7-8cee0956d990-kube-api-access-6cxn7\") on node \"crc\" DevicePath \"\"" Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.949691 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" event={"ID":"a71c86d5-acae-401d-bea7-8cee0956d990","Type":"ContainerDied","Data":"06dd2c0b14682d57a9d4c3f2399b2b8728a168b1231ab47351c72b21120c51db"} Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.949752 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06dd2c0b14682d57a9d4c3f2399b2b8728a168b1231ab47351c72b21120c51db" Oct 04 08:45:17 crc kubenswrapper[4969]: I1004 08:45:17.949780 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.045643 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2"] Oct 04 08:45:18 crc kubenswrapper[4969]: E1004 08:45:18.046031 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69cdc4bb-3b8e-4094-9ee6-fab85d869ffa" containerName="collect-profiles" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.046048 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="69cdc4bb-3b8e-4094-9ee6-fab85d869ffa" containerName="collect-profiles" Oct 04 08:45:18 crc kubenswrapper[4969]: E1004 08:45:18.046075 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a71c86d5-acae-401d-bea7-8cee0956d990" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.046082 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a71c86d5-acae-401d-bea7-8cee0956d990" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.046275 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="69cdc4bb-3b8e-4094-9ee6-fab85d869ffa" containerName="collect-profiles" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.046302 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="a71c86d5-acae-401d-bea7-8cee0956d990" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.047041 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.048997 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.050678 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.050711 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.050836 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.065771 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2"] Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.151460 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.151634 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sbrp\" (UniqueName: \"kubernetes.io/projected/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-kube-api-access-5sbrp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.152037 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.253906 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.254252 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.254610 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sbrp\" (UniqueName: \"kubernetes.io/projected/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-kube-api-access-5sbrp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.259569 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.264553 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.279281 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sbrp\" (UniqueName: \"kubernetes.io/projected/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-kube-api-access-5sbrp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.365260 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.885651 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2"] Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.885911 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 08:45:18 crc kubenswrapper[4969]: I1004 08:45:18.961574 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" event={"ID":"48e0d7b9-c122-4cee-84b3-1e89a72af2ad","Type":"ContainerStarted","Data":"06c595bcf4d88078209997c7687f6f9a6cd33da2e8a6c3dcffb6c0fa04b137bf"} Oct 04 08:45:19 crc kubenswrapper[4969]: I1004 08:45:19.971507 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" event={"ID":"48e0d7b9-c122-4cee-84b3-1e89a72af2ad","Type":"ContainerStarted","Data":"b50766e16ea5fadd0ffbca02ba67ae9f05b47a765c291cc294150b23a584e0e0"} Oct 04 08:45:19 crc kubenswrapper[4969]: I1004 08:45:19.995264 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" podStartSLOduration=1.4681977210000001 podStartE2EDuration="1.995240757s" podCreationTimestamp="2025-10-04 08:45:18 +0000 UTC" firstStartedPulling="2025-10-04 08:45:18.885664906 +0000 UTC m=+1746.639933720" lastFinishedPulling="2025-10-04 08:45:19.412707942 +0000 UTC m=+1747.166976756" observedRunningTime="2025-10-04 08:45:19.985933712 +0000 UTC m=+1747.740202536" watchObservedRunningTime="2025-10-04 08:45:19.995240757 +0000 UTC m=+1747.749509581" Oct 04 08:45:24 crc kubenswrapper[4969]: I1004 08:45:24.056027 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:45:24 crc kubenswrapper[4969]: E1004 08:45:24.057022 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:45:26 crc kubenswrapper[4969]: I1004 08:45:26.037053 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-cf9ls"] Oct 04 08:45:26 crc kubenswrapper[4969]: I1004 08:45:26.047065 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-cf9ls"] Oct 04 08:45:27 crc kubenswrapper[4969]: I1004 08:45:27.078405 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2861dd9f-5291-4f2a-bb30-0b9c54baa5cf" path="/var/lib/kubelet/pods/2861dd9f-5291-4f2a-bb30-0b9c54baa5cf/volumes" Oct 04 08:45:29 crc kubenswrapper[4969]: I1004 08:45:29.086336 4969 scope.go:117] "RemoveContainer" containerID="5010c1e44d7549010cfe78af5a5c7d1d37af909ead2271f81751d7689e42402b" Oct 04 08:45:29 crc kubenswrapper[4969]: I1004 08:45:29.123513 4969 scope.go:117] "RemoveContainer" containerID="4125e33fb697f71057aa46cf79725094aaf92fada934f2163ea811c16b4176ea" Oct 04 08:45:29 crc kubenswrapper[4969]: I1004 08:45:29.165483 4969 scope.go:117] "RemoveContainer" containerID="f22e41ad155ab7a114bc1d62a5a37050455b6bf0e9177b4657828cec48d2a144" Oct 04 08:45:29 crc kubenswrapper[4969]: I1004 08:45:29.218928 4969 scope.go:117] "RemoveContainer" containerID="fa7d3b44e484120a5a3ea50089d66bad7288ad199d49aca435555dc55b054f41" Oct 04 08:45:30 crc kubenswrapper[4969]: I1004 08:45:30.028005 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-7fphd"] Oct 04 08:45:30 crc kubenswrapper[4969]: I1004 08:45:30.037656 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-7fphd"] Oct 04 08:45:30 crc kubenswrapper[4969]: I1004 08:45:30.047733 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-bx4tb"] Oct 04 08:45:30 crc kubenswrapper[4969]: I1004 08:45:30.055155 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-bx4tb"] Oct 04 08:45:31 crc kubenswrapper[4969]: I1004 08:45:31.041712 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-wjg5p"] Oct 04 08:45:31 crc kubenswrapper[4969]: I1004 08:45:31.053680 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-wjg5p"] Oct 04 08:45:31 crc kubenswrapper[4969]: I1004 08:45:31.074321 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ff3f16f-8103-49a7-8a5c-07b6b7f399e7" path="/var/lib/kubelet/pods/1ff3f16f-8103-49a7-8a5c-07b6b7f399e7/volumes" Oct 04 08:45:31 crc kubenswrapper[4969]: I1004 08:45:31.075310 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46a217c4-ea44-4608-9300-65410bec2b88" path="/var/lib/kubelet/pods/46a217c4-ea44-4608-9300-65410bec2b88/volumes" Oct 04 08:45:31 crc kubenswrapper[4969]: I1004 08:45:31.076016 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f" path="/var/lib/kubelet/pods/f3ee3d2e-3c10-4bda-8a4f-435d0fc8a16f/volumes" Oct 04 08:45:36 crc kubenswrapper[4969]: I1004 08:45:36.055855 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:45:36 crc kubenswrapper[4969]: E1004 08:45:36.056934 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:45:37 crc kubenswrapper[4969]: I1004 08:45:37.054163 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-xjc7p"] Oct 04 08:45:37 crc kubenswrapper[4969]: I1004 08:45:37.070481 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-xjc7p"] Oct 04 08:45:39 crc kubenswrapper[4969]: I1004 08:45:39.087702 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="278ebf83-6c66-4d97-8b5c-a7a83ace00b7" path="/var/lib/kubelet/pods/278ebf83-6c66-4d97-8b5c-a7a83ace00b7/volumes" Oct 04 08:45:42 crc kubenswrapper[4969]: I1004 08:45:42.032044 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-74db-account-create-h78hj"] Oct 04 08:45:42 crc kubenswrapper[4969]: I1004 08:45:42.048703 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-b89e-account-create-zlbmg"] Oct 04 08:45:42 crc kubenswrapper[4969]: I1004 08:45:42.058851 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-74db-account-create-h78hj"] Oct 04 08:45:42 crc kubenswrapper[4969]: I1004 08:45:42.067835 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-b89e-account-create-zlbmg"] Oct 04 08:45:43 crc kubenswrapper[4969]: I1004 08:45:43.035461 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-56df-account-create-8jbkc"] Oct 04 08:45:43 crc kubenswrapper[4969]: I1004 08:45:43.072682 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a41b61-3a97-4d3c-94b6-f76f5d58abe4" path="/var/lib/kubelet/pods/00a41b61-3a97-4d3c-94b6-f76f5d58abe4/volumes" Oct 04 08:45:43 crc kubenswrapper[4969]: I1004 08:45:43.073487 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4af2ba9-d233-4928-9a54-b5bf58ee50a9" path="/var/lib/kubelet/pods/d4af2ba9-d233-4928-9a54-b5bf58ee50a9/volumes" Oct 04 08:45:43 crc kubenswrapper[4969]: I1004 08:45:43.074306 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-56df-account-create-8jbkc"] Oct 04 08:45:45 crc kubenswrapper[4969]: I1004 08:45:45.075320 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16495ba9-75b1-471b-8152-e477f327d94f" path="/var/lib/kubelet/pods/16495ba9-75b1-471b-8152-e477f327d94f/volumes" Oct 04 08:45:48 crc kubenswrapper[4969]: I1004 08:45:48.056060 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:45:48 crc kubenswrapper[4969]: E1004 08:45:48.056621 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:46:03 crc kubenswrapper[4969]: I1004 08:46:03.061079 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:46:03 crc kubenswrapper[4969]: E1004 08:46:03.061693 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:46:15 crc kubenswrapper[4969]: I1004 08:46:15.056475 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:46:15 crc kubenswrapper[4969]: E1004 08:46:15.057381 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:46:18 crc kubenswrapper[4969]: I1004 08:46:18.076389 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjwv2"] Oct 04 08:46:18 crc kubenswrapper[4969]: I1004 08:46:18.086167 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjwv2"] Oct 04 08:46:19 crc kubenswrapper[4969]: I1004 08:46:19.069561 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a0efc12-8421-489a-8fe9-97ae54119cc8" path="/var/lib/kubelet/pods/8a0efc12-8421-489a-8fe9-97ae54119cc8/volumes" Oct 04 08:46:29 crc kubenswrapper[4969]: I1004 08:46:29.056287 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:46:29 crc kubenswrapper[4969]: E1004 08:46:29.057406 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:46:29 crc kubenswrapper[4969]: I1004 08:46:29.421495 4969 scope.go:117] "RemoveContainer" containerID="51525ee67e0a0f9a69dc004390b4adfb3a6e5bdcf597d1a88687b00712ab9cc7" Oct 04 08:46:29 crc kubenswrapper[4969]: I1004 08:46:29.449387 4969 scope.go:117] "RemoveContainer" containerID="02584b220d350e940b2c33e9b081a7f4c3cceec5d5da78988633c34cfd0a3eab" Oct 04 08:46:29 crc kubenswrapper[4969]: I1004 08:46:29.491217 4969 scope.go:117] "RemoveContainer" containerID="56d1c29e6a36964868664598f9c3ae20edc0b600866ce852fdb6366ee59594f8" Oct 04 08:46:29 crc kubenswrapper[4969]: I1004 08:46:29.560715 4969 scope.go:117] "RemoveContainer" containerID="65ea1e18cdd39af7f4a6148bf32f5950ef0fd483cbc159d23dca96f64aecff13" Oct 04 08:46:29 crc kubenswrapper[4969]: I1004 08:46:29.639607 4969 scope.go:117] "RemoveContainer" containerID="ca6768d0d1b8429421ce74128dc645f9edaaf7b04b4952878c75bdd98d883fc3" Oct 04 08:46:29 crc kubenswrapper[4969]: I1004 08:46:29.712827 4969 scope.go:117] "RemoveContainer" containerID="89fbd59ae652ba88a603944657850a84aafc6ce9155b6c8fee9714ebe53fa82e" Oct 04 08:46:29 crc kubenswrapper[4969]: I1004 08:46:29.744752 4969 scope.go:117] "RemoveContainer" containerID="449fe9b996875ca6746e78842bfb0e9eefbf4e3ad978360a06766e20c54204a8" Oct 04 08:46:29 crc kubenswrapper[4969]: I1004 08:46:29.795873 4969 scope.go:117] "RemoveContainer" containerID="8e2e3e952189d7105e232645f1ca31ee92db17c6482688944a3b1999f64fc3fd" Oct 04 08:46:36 crc kubenswrapper[4969]: I1004 08:46:36.797852 4969 generic.go:334] "Generic (PLEG): container finished" podID="48e0d7b9-c122-4cee-84b3-1e89a72af2ad" containerID="b50766e16ea5fadd0ffbca02ba67ae9f05b47a765c291cc294150b23a584e0e0" exitCode=0 Oct 04 08:46:36 crc kubenswrapper[4969]: I1004 08:46:36.797933 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" event={"ID":"48e0d7b9-c122-4cee-84b3-1e89a72af2ad","Type":"ContainerDied","Data":"b50766e16ea5fadd0ffbca02ba67ae9f05b47a765c291cc294150b23a584e0e0"} Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.266205 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.388904 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-inventory\") pod \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.389077 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-ssh-key\") pod \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.389179 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sbrp\" (UniqueName: \"kubernetes.io/projected/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-kube-api-access-5sbrp\") pod \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\" (UID: \"48e0d7b9-c122-4cee-84b3-1e89a72af2ad\") " Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.395665 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-kube-api-access-5sbrp" (OuterVolumeSpecName: "kube-api-access-5sbrp") pod "48e0d7b9-c122-4cee-84b3-1e89a72af2ad" (UID: "48e0d7b9-c122-4cee-84b3-1e89a72af2ad"). InnerVolumeSpecName "kube-api-access-5sbrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.439624 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-inventory" (OuterVolumeSpecName: "inventory") pod "48e0d7b9-c122-4cee-84b3-1e89a72af2ad" (UID: "48e0d7b9-c122-4cee-84b3-1e89a72af2ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.442924 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48e0d7b9-c122-4cee-84b3-1e89a72af2ad" (UID: "48e0d7b9-c122-4cee-84b3-1e89a72af2ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.498593 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.498649 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.498672 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sbrp\" (UniqueName: \"kubernetes.io/projected/48e0d7b9-c122-4cee-84b3-1e89a72af2ad-kube-api-access-5sbrp\") on node \"crc\" DevicePath \"\"" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.826807 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" event={"ID":"48e0d7b9-c122-4cee-84b3-1e89a72af2ad","Type":"ContainerDied","Data":"06c595bcf4d88078209997c7687f6f9a6cd33da2e8a6c3dcffb6c0fa04b137bf"} Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.827185 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06c595bcf4d88078209997c7687f6f9a6cd33da2e8a6c3dcffb6c0fa04b137bf" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.827022 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.914596 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7"] Oct 04 08:46:38 crc kubenswrapper[4969]: E1004 08:46:38.915128 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e0d7b9-c122-4cee-84b3-1e89a72af2ad" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.915160 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e0d7b9-c122-4cee-84b3-1e89a72af2ad" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.915476 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e0d7b9-c122-4cee-84b3-1e89a72af2ad" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.916499 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.921062 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.921374 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.921629 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.921839 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:46:38 crc kubenswrapper[4969]: I1004 08:46:38.930150 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7"] Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.007732 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.007810 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.007851 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb7tm\" (UniqueName: \"kubernetes.io/projected/8c7813c0-2dcd-4d30-9f8c-fac53077417f-kube-api-access-lb7tm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.110556 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.110643 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.110683 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb7tm\" (UniqueName: \"kubernetes.io/projected/8c7813c0-2dcd-4d30-9f8c-fac53077417f-kube-api-access-lb7tm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.115796 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.117741 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.129705 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb7tm\" (UniqueName: \"kubernetes.io/projected/8c7813c0-2dcd-4d30-9f8c-fac53077417f-kube-api-access-lb7tm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.276914 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:39 crc kubenswrapper[4969]: I1004 08:46:39.879888 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7"] Oct 04 08:46:39 crc kubenswrapper[4969]: W1004 08:46:39.887001 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c7813c0_2dcd_4d30_9f8c_fac53077417f.slice/crio-940de605b5f5eb8afe642a8360231af6d8f2a97f88806355733298c4e5424c8b WatchSource:0}: Error finding container 940de605b5f5eb8afe642a8360231af6d8f2a97f88806355733298c4e5424c8b: Status 404 returned error can't find the container with id 940de605b5f5eb8afe642a8360231af6d8f2a97f88806355733298c4e5424c8b Oct 04 08:46:40 crc kubenswrapper[4969]: I1004 08:46:40.054788 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:46:40 crc kubenswrapper[4969]: E1004 08:46:40.055258 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:46:40 crc kubenswrapper[4969]: I1004 08:46:40.847893 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" event={"ID":"8c7813c0-2dcd-4d30-9f8c-fac53077417f","Type":"ContainerStarted","Data":"940de605b5f5eb8afe642a8360231af6d8f2a97f88806355733298c4e5424c8b"} Oct 04 08:46:42 crc kubenswrapper[4969]: I1004 08:46:42.874798 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" event={"ID":"8c7813c0-2dcd-4d30-9f8c-fac53077417f","Type":"ContainerStarted","Data":"9f4c7f4811cbea72bfd9f3417b44d31ba3867b5b2b293284bf4ac4f8eb95373f"} Oct 04 08:46:42 crc kubenswrapper[4969]: I1004 08:46:42.915146 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" podStartSLOduration=3.355500311 podStartE2EDuration="4.915123212s" podCreationTimestamp="2025-10-04 08:46:38 +0000 UTC" firstStartedPulling="2025-10-04 08:46:39.889661971 +0000 UTC m=+1827.643930795" lastFinishedPulling="2025-10-04 08:46:41.449284872 +0000 UTC m=+1829.203553696" observedRunningTime="2025-10-04 08:46:42.901289943 +0000 UTC m=+1830.655558777" watchObservedRunningTime="2025-10-04 08:46:42.915123212 +0000 UTC m=+1830.669392046" Oct 04 08:46:43 crc kubenswrapper[4969]: I1004 08:46:43.073050 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-9hbl9"] Oct 04 08:46:43 crc kubenswrapper[4969]: I1004 08:46:43.079758 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-9hbl9"] Oct 04 08:46:45 crc kubenswrapper[4969]: I1004 08:46:45.072379 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cebc3b1f-5a32-4255-ae14-4fea09ad9f12" path="/var/lib/kubelet/pods/cebc3b1f-5a32-4255-ae14-4fea09ad9f12/volumes" Oct 04 08:46:47 crc kubenswrapper[4969]: I1004 08:46:47.044315 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9xhfr"] Oct 04 08:46:47 crc kubenswrapper[4969]: I1004 08:46:47.069651 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9xhfr"] Oct 04 08:46:48 crc kubenswrapper[4969]: I1004 08:46:48.955501 4969 generic.go:334] "Generic (PLEG): container finished" podID="8c7813c0-2dcd-4d30-9f8c-fac53077417f" containerID="9f4c7f4811cbea72bfd9f3417b44d31ba3867b5b2b293284bf4ac4f8eb95373f" exitCode=0 Oct 04 08:46:48 crc kubenswrapper[4969]: I1004 08:46:48.955584 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" event={"ID":"8c7813c0-2dcd-4d30-9f8c-fac53077417f","Type":"ContainerDied","Data":"9f4c7f4811cbea72bfd9f3417b44d31ba3867b5b2b293284bf4ac4f8eb95373f"} Oct 04 08:46:49 crc kubenswrapper[4969]: I1004 08:46:49.071092 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2160e9f1-8d10-4ddf-931f-972f80fed6ba" path="/var/lib/kubelet/pods/2160e9f1-8d10-4ddf-931f-972f80fed6ba/volumes" Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.514165 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.597604 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-ssh-key\") pod \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.598129 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-inventory\") pod \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.598194 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb7tm\" (UniqueName: \"kubernetes.io/projected/8c7813c0-2dcd-4d30-9f8c-fac53077417f-kube-api-access-lb7tm\") pod \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\" (UID: \"8c7813c0-2dcd-4d30-9f8c-fac53077417f\") " Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.605661 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c7813c0-2dcd-4d30-9f8c-fac53077417f-kube-api-access-lb7tm" (OuterVolumeSpecName: "kube-api-access-lb7tm") pod "8c7813c0-2dcd-4d30-9f8c-fac53077417f" (UID: "8c7813c0-2dcd-4d30-9f8c-fac53077417f"). InnerVolumeSpecName "kube-api-access-lb7tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.628406 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8c7813c0-2dcd-4d30-9f8c-fac53077417f" (UID: "8c7813c0-2dcd-4d30-9f8c-fac53077417f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.644986 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-inventory" (OuterVolumeSpecName: "inventory") pod "8c7813c0-2dcd-4d30-9f8c-fac53077417f" (UID: "8c7813c0-2dcd-4d30-9f8c-fac53077417f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.701167 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.701206 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb7tm\" (UniqueName: \"kubernetes.io/projected/8c7813c0-2dcd-4d30-9f8c-fac53077417f-kube-api-access-lb7tm\") on node \"crc\" DevicePath \"\"" Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.701218 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c7813c0-2dcd-4d30-9f8c-fac53077417f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.984897 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" event={"ID":"8c7813c0-2dcd-4d30-9f8c-fac53077417f","Type":"ContainerDied","Data":"940de605b5f5eb8afe642a8360231af6d8f2a97f88806355733298c4e5424c8b"} Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.985150 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="940de605b5f5eb8afe642a8360231af6d8f2a97f88806355733298c4e5424c8b" Oct 04 08:46:50 crc kubenswrapper[4969]: I1004 08:46:50.985232 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.105297 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6"] Oct 04 08:46:51 crc kubenswrapper[4969]: E1004 08:46:51.105730 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7813c0-2dcd-4d30-9f8c-fac53077417f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.105752 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7813c0-2dcd-4d30-9f8c-fac53077417f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.105956 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7813c0-2dcd-4d30-9f8c-fac53077417f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.106598 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6"] Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.106690 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.109002 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.109167 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.109353 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.110470 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.210614 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cqqr6\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.210850 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5jqg\" (UniqueName: \"kubernetes.io/projected/579fed39-17e4-412c-b76b-139db1b5938e-kube-api-access-r5jqg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cqqr6\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.211003 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cqqr6\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.312735 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cqqr6\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.312936 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5jqg\" (UniqueName: \"kubernetes.io/projected/579fed39-17e4-412c-b76b-139db1b5938e-kube-api-access-r5jqg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cqqr6\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.312981 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cqqr6\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.317390 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cqqr6\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.320286 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cqqr6\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.332228 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5jqg\" (UniqueName: \"kubernetes.io/projected/579fed39-17e4-412c-b76b-139db1b5938e-kube-api-access-r5jqg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cqqr6\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.430068 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:46:51 crc kubenswrapper[4969]: I1004 08:46:51.986770 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6"] Oct 04 08:46:52 crc kubenswrapper[4969]: I1004 08:46:52.055411 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:46:52 crc kubenswrapper[4969]: E1004 08:46:52.056135 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:46:53 crc kubenswrapper[4969]: I1004 08:46:53.010198 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" event={"ID":"579fed39-17e4-412c-b76b-139db1b5938e","Type":"ContainerStarted","Data":"f5a9b59f32b1d24ab1022749cc9ddbd27385005feeed89a21f836fb1c1effd30"} Oct 04 08:46:54 crc kubenswrapper[4969]: I1004 08:46:54.021722 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" event={"ID":"579fed39-17e4-412c-b76b-139db1b5938e","Type":"ContainerStarted","Data":"877738c42d70b817e6b27a6cd8b820f640d871a8d8fc3e54b197b9a6d9ebef8b"} Oct 04 08:46:54 crc kubenswrapper[4969]: I1004 08:46:54.052304 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" podStartSLOduration=2.077643712 podStartE2EDuration="3.052279511s" podCreationTimestamp="2025-10-04 08:46:51 +0000 UTC" firstStartedPulling="2025-10-04 08:46:51.999220506 +0000 UTC m=+1839.753489350" lastFinishedPulling="2025-10-04 08:46:52.973856325 +0000 UTC m=+1840.728125149" observedRunningTime="2025-10-04 08:46:54.04325453 +0000 UTC m=+1841.797523384" watchObservedRunningTime="2025-10-04 08:46:54.052279511 +0000 UTC m=+1841.806548335" Oct 04 08:47:05 crc kubenswrapper[4969]: I1004 08:47:05.056057 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:47:05 crc kubenswrapper[4969]: E1004 08:47:05.057316 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:47:19 crc kubenswrapper[4969]: I1004 08:47:19.056124 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:47:19 crc kubenswrapper[4969]: E1004 08:47:19.057290 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:47:28 crc kubenswrapper[4969]: I1004 08:47:28.051043 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-tj6f9"] Oct 04 08:47:28 crc kubenswrapper[4969]: I1004 08:47:28.058632 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-tj6f9"] Oct 04 08:47:29 crc kubenswrapper[4969]: I1004 08:47:29.075798 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f7bb15e-5c1f-4811-acc0-421d2bd13d88" path="/var/lib/kubelet/pods/1f7bb15e-5c1f-4811-acc0-421d2bd13d88/volumes" Oct 04 08:47:30 crc kubenswrapper[4969]: I1004 08:47:30.226617 4969 scope.go:117] "RemoveContainer" containerID="e7ff900bdfe9aa8996d2b493e7f1d534772119ba64f3a4f72ccf3a12f38e9ea7" Oct 04 08:47:30 crc kubenswrapper[4969]: I1004 08:47:30.314099 4969 scope.go:117] "RemoveContainer" containerID="906c11d6be4da220c5de9e0d3c69f0a215ef98e2509527008b4d4ab455a32467" Oct 04 08:47:30 crc kubenswrapper[4969]: I1004 08:47:30.359046 4969 scope.go:117] "RemoveContainer" containerID="5c3ecbc26a9d98cfe05688d490ec149ed5852e06bd1bf98f08311d10f1aef34a" Oct 04 08:47:32 crc kubenswrapper[4969]: I1004 08:47:32.055800 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:47:32 crc kubenswrapper[4969]: E1004 08:47:32.056461 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:47:34 crc kubenswrapper[4969]: I1004 08:47:34.511970 4969 generic.go:334] "Generic (PLEG): container finished" podID="579fed39-17e4-412c-b76b-139db1b5938e" containerID="877738c42d70b817e6b27a6cd8b820f640d871a8d8fc3e54b197b9a6d9ebef8b" exitCode=0 Oct 04 08:47:34 crc kubenswrapper[4969]: I1004 08:47:34.512059 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" event={"ID":"579fed39-17e4-412c-b76b-139db1b5938e","Type":"ContainerDied","Data":"877738c42d70b817e6b27a6cd8b820f640d871a8d8fc3e54b197b9a6d9ebef8b"} Oct 04 08:47:35 crc kubenswrapper[4969]: I1004 08:47:35.923084 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.093296 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5jqg\" (UniqueName: \"kubernetes.io/projected/579fed39-17e4-412c-b76b-139db1b5938e-kube-api-access-r5jqg\") pod \"579fed39-17e4-412c-b76b-139db1b5938e\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.093387 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-inventory\") pod \"579fed39-17e4-412c-b76b-139db1b5938e\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.093505 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-ssh-key\") pod \"579fed39-17e4-412c-b76b-139db1b5938e\" (UID: \"579fed39-17e4-412c-b76b-139db1b5938e\") " Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.101046 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/579fed39-17e4-412c-b76b-139db1b5938e-kube-api-access-r5jqg" (OuterVolumeSpecName: "kube-api-access-r5jqg") pod "579fed39-17e4-412c-b76b-139db1b5938e" (UID: "579fed39-17e4-412c-b76b-139db1b5938e"). InnerVolumeSpecName "kube-api-access-r5jqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.123157 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "579fed39-17e4-412c-b76b-139db1b5938e" (UID: "579fed39-17e4-412c-b76b-139db1b5938e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.144458 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-inventory" (OuterVolumeSpecName: "inventory") pod "579fed39-17e4-412c-b76b-139db1b5938e" (UID: "579fed39-17e4-412c-b76b-139db1b5938e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.195715 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5jqg\" (UniqueName: \"kubernetes.io/projected/579fed39-17e4-412c-b76b-139db1b5938e-kube-api-access-r5jqg\") on node \"crc\" DevicePath \"\"" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.195754 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.195763 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579fed39-17e4-412c-b76b-139db1b5938e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.534577 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" event={"ID":"579fed39-17e4-412c-b76b-139db1b5938e","Type":"ContainerDied","Data":"f5a9b59f32b1d24ab1022749cc9ddbd27385005feeed89a21f836fb1c1effd30"} Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.534653 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5a9b59f32b1d24ab1022749cc9ddbd27385005feeed89a21f836fb1c1effd30" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.534662 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cqqr6" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.653391 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk"] Oct 04 08:47:36 crc kubenswrapper[4969]: E1004 08:47:36.653951 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="579fed39-17e4-412c-b76b-139db1b5938e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.653982 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="579fed39-17e4-412c-b76b-139db1b5938e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.654236 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="579fed39-17e4-412c-b76b-139db1b5938e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.655125 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.657100 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.657314 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.657952 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.661610 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.664727 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk"] Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.808323 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.808530 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.808617 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4jjk\" (UniqueName: \"kubernetes.io/projected/85fdbfa3-25a5-4a92-9473-198d3cf416c8-kube-api-access-p4jjk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.910186 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.910242 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4jjk\" (UniqueName: \"kubernetes.io/projected/85fdbfa3-25a5-4a92-9473-198d3cf416c8-kube-api-access-p4jjk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.910377 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.916000 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.921132 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.937552 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4jjk\" (UniqueName: \"kubernetes.io/projected/85fdbfa3-25a5-4a92-9473-198d3cf416c8-kube-api-access-p4jjk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:36 crc kubenswrapper[4969]: I1004 08:47:36.983096 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:47:37 crc kubenswrapper[4969]: I1004 08:47:37.540460 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk"] Oct 04 08:47:38 crc kubenswrapper[4969]: I1004 08:47:38.555967 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" event={"ID":"85fdbfa3-25a5-4a92-9473-198d3cf416c8","Type":"ContainerStarted","Data":"8d936ee1b1f320802ca56b310bfe125f6b09081d1676008c7ae409f794b4d50f"} Oct 04 08:47:39 crc kubenswrapper[4969]: I1004 08:47:39.567464 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" event={"ID":"85fdbfa3-25a5-4a92-9473-198d3cf416c8","Type":"ContainerStarted","Data":"c776e94240724a8f665312749bce0d5b2858250b233d117c1db444f111c46afe"} Oct 04 08:47:45 crc kubenswrapper[4969]: I1004 08:47:45.055085 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:47:45 crc kubenswrapper[4969]: E1004 08:47:45.056080 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:47:56 crc kubenswrapper[4969]: I1004 08:47:56.055741 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:47:56 crc kubenswrapper[4969]: I1004 08:47:56.739535 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"f87372b3dbcc42c8848b2ff115cee7651dab7a8ca5bb7cad720f3bfdc85a0f10"} Oct 04 08:47:56 crc kubenswrapper[4969]: I1004 08:47:56.764348 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" podStartSLOduration=19.283059282 podStartE2EDuration="20.764321533s" podCreationTimestamp="2025-10-04 08:47:36 +0000 UTC" firstStartedPulling="2025-10-04 08:47:37.551956024 +0000 UTC m=+1885.306224828" lastFinishedPulling="2025-10-04 08:47:39.033218255 +0000 UTC m=+1886.787487079" observedRunningTime="2025-10-04 08:47:39.597641374 +0000 UTC m=+1887.351910228" watchObservedRunningTime="2025-10-04 08:47:56.764321533 +0000 UTC m=+1904.518590347" Oct 04 08:48:38 crc kubenswrapper[4969]: I1004 08:48:38.148651 4969 generic.go:334] "Generic (PLEG): container finished" podID="85fdbfa3-25a5-4a92-9473-198d3cf416c8" containerID="c776e94240724a8f665312749bce0d5b2858250b233d117c1db444f111c46afe" exitCode=2 Oct 04 08:48:38 crc kubenswrapper[4969]: I1004 08:48:38.148770 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" event={"ID":"85fdbfa3-25a5-4a92-9473-198d3cf416c8","Type":"ContainerDied","Data":"c776e94240724a8f665312749bce0d5b2858250b233d117c1db444f111c46afe"} Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.640000 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.738261 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4jjk\" (UniqueName: \"kubernetes.io/projected/85fdbfa3-25a5-4a92-9473-198d3cf416c8-kube-api-access-p4jjk\") pod \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.738355 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-inventory\") pod \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.738610 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-ssh-key\") pod \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\" (UID: \"85fdbfa3-25a5-4a92-9473-198d3cf416c8\") " Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.762225 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85fdbfa3-25a5-4a92-9473-198d3cf416c8-kube-api-access-p4jjk" (OuterVolumeSpecName: "kube-api-access-p4jjk") pod "85fdbfa3-25a5-4a92-9473-198d3cf416c8" (UID: "85fdbfa3-25a5-4a92-9473-198d3cf416c8"). InnerVolumeSpecName "kube-api-access-p4jjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.779546 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "85fdbfa3-25a5-4a92-9473-198d3cf416c8" (UID: "85fdbfa3-25a5-4a92-9473-198d3cf416c8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.795928 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-inventory" (OuterVolumeSpecName: "inventory") pod "85fdbfa3-25a5-4a92-9473-198d3cf416c8" (UID: "85fdbfa3-25a5-4a92-9473-198d3cf416c8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.841160 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4jjk\" (UniqueName: \"kubernetes.io/projected/85fdbfa3-25a5-4a92-9473-198d3cf416c8-kube-api-access-p4jjk\") on node \"crc\" DevicePath \"\"" Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.841204 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:48:39 crc kubenswrapper[4969]: I1004 08:48:39.841216 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85fdbfa3-25a5-4a92-9473-198d3cf416c8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:48:40 crc kubenswrapper[4969]: I1004 08:48:40.170412 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" event={"ID":"85fdbfa3-25a5-4a92-9473-198d3cf416c8","Type":"ContainerDied","Data":"8d936ee1b1f320802ca56b310bfe125f6b09081d1676008c7ae409f794b4d50f"} Oct 04 08:48:40 crc kubenswrapper[4969]: I1004 08:48:40.170798 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d936ee1b1f320802ca56b310bfe125f6b09081d1676008c7ae409f794b4d50f" Oct 04 08:48:40 crc kubenswrapper[4969]: I1004 08:48:40.170878 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.043972 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n"] Oct 04 08:48:47 crc kubenswrapper[4969]: E1004 08:48:47.045398 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85fdbfa3-25a5-4a92-9473-198d3cf416c8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.045472 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="85fdbfa3-25a5-4a92-9473-198d3cf416c8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.045892 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="85fdbfa3-25a5-4a92-9473-198d3cf416c8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.047014 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.051579 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.051714 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.052217 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.054026 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.094865 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t796n\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.095222 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbmfk\" (UniqueName: \"kubernetes.io/projected/19b4ae96-fa76-4d96-b4e0-d0058345fc23-kube-api-access-kbmfk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t796n\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.095471 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t796n\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.100143 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n"] Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.199148 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t796n\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.199534 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbmfk\" (UniqueName: \"kubernetes.io/projected/19b4ae96-fa76-4d96-b4e0-d0058345fc23-kube-api-access-kbmfk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t796n\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.199701 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t796n\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.209201 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t796n\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.211927 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t796n\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.232757 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbmfk\" (UniqueName: \"kubernetes.io/projected/19b4ae96-fa76-4d96-b4e0-d0058345fc23-kube-api-access-kbmfk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t796n\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:47 crc kubenswrapper[4969]: I1004 08:48:47.389237 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:48:48 crc kubenswrapper[4969]: I1004 08:48:48.025876 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n"] Oct 04 08:48:48 crc kubenswrapper[4969]: I1004 08:48:48.257047 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" event={"ID":"19b4ae96-fa76-4d96-b4e0-d0058345fc23","Type":"ContainerStarted","Data":"8827fe252d97834df4c7d81a68be03118847cdfd78c4ea45aa4215c1911d6bf1"} Oct 04 08:48:49 crc kubenswrapper[4969]: I1004 08:48:49.269362 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" event={"ID":"19b4ae96-fa76-4d96-b4e0-d0058345fc23","Type":"ContainerStarted","Data":"79bee0fd376c80a3f61b36e0d2cf469dc7b5771fb1fe716cf332b0acc3a31478"} Oct 04 08:48:49 crc kubenswrapper[4969]: I1004 08:48:49.288530 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" podStartSLOduration=1.492835103 podStartE2EDuration="2.288509107s" podCreationTimestamp="2025-10-04 08:48:47 +0000 UTC" firstStartedPulling="2025-10-04 08:48:48.038886871 +0000 UTC m=+1955.793155695" lastFinishedPulling="2025-10-04 08:48:48.834560885 +0000 UTC m=+1956.588829699" observedRunningTime="2025-10-04 08:48:49.285824461 +0000 UTC m=+1957.040093305" watchObservedRunningTime="2025-10-04 08:48:49.288509107 +0000 UTC m=+1957.042777921" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.174724 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nwfcs"] Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.178515 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.218485 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nwfcs"] Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.270603 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-catalog-content\") pod \"certified-operators-nwfcs\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.270687 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-utilities\") pod \"certified-operators-nwfcs\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.270731 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdmsc\" (UniqueName: \"kubernetes.io/projected/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-kube-api-access-cdmsc\") pod \"certified-operators-nwfcs\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.372865 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-catalog-content\") pod \"certified-operators-nwfcs\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.372943 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-utilities\") pod \"certified-operators-nwfcs\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.372987 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdmsc\" (UniqueName: \"kubernetes.io/projected/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-kube-api-access-cdmsc\") pod \"certified-operators-nwfcs\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.373537 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-catalog-content\") pod \"certified-operators-nwfcs\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.373606 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-utilities\") pod \"certified-operators-nwfcs\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.405932 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdmsc\" (UniqueName: \"kubernetes.io/projected/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-kube-api-access-cdmsc\") pod \"certified-operators-nwfcs\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:04 crc kubenswrapper[4969]: I1004 08:49:04.507775 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:05 crc kubenswrapper[4969]: I1004 08:49:05.092502 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nwfcs"] Oct 04 08:49:05 crc kubenswrapper[4969]: I1004 08:49:05.444473 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwfcs" event={"ID":"21d233b3-8a4c-4f3d-b89f-40af9b5ca381","Type":"ContainerStarted","Data":"1f554cd17f7d36ae623ed82b7a05641912a522c40e3fd8c27819dbba3d7d3b17"} Oct 04 08:49:05 crc kubenswrapper[4969]: I1004 08:49:05.444515 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwfcs" event={"ID":"21d233b3-8a4c-4f3d-b89f-40af9b5ca381","Type":"ContainerStarted","Data":"8ec7a8b87adb030ddb0115ecdbf50b933ba45a301a567a92b21db46ac187d181"} Oct 04 08:49:06 crc kubenswrapper[4969]: I1004 08:49:06.461183 4969 generic.go:334] "Generic (PLEG): container finished" podID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerID="1f554cd17f7d36ae623ed82b7a05641912a522c40e3fd8c27819dbba3d7d3b17" exitCode=0 Oct 04 08:49:06 crc kubenswrapper[4969]: I1004 08:49:06.461265 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwfcs" event={"ID":"21d233b3-8a4c-4f3d-b89f-40af9b5ca381","Type":"ContainerDied","Data":"1f554cd17f7d36ae623ed82b7a05641912a522c40e3fd8c27819dbba3d7d3b17"} Oct 04 08:49:08 crc kubenswrapper[4969]: I1004 08:49:08.479530 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwfcs" event={"ID":"21d233b3-8a4c-4f3d-b89f-40af9b5ca381","Type":"ContainerStarted","Data":"07c68e0744c7c315a32c00f3639eabc232d85292a2af9fc167b6933d751fc3e4"} Oct 04 08:49:09 crc kubenswrapper[4969]: I1004 08:49:09.490369 4969 generic.go:334] "Generic (PLEG): container finished" podID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerID="07c68e0744c7c315a32c00f3639eabc232d85292a2af9fc167b6933d751fc3e4" exitCode=0 Oct 04 08:49:09 crc kubenswrapper[4969]: I1004 08:49:09.490422 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwfcs" event={"ID":"21d233b3-8a4c-4f3d-b89f-40af9b5ca381","Type":"ContainerDied","Data":"07c68e0744c7c315a32c00f3639eabc232d85292a2af9fc167b6933d751fc3e4"} Oct 04 08:49:11 crc kubenswrapper[4969]: I1004 08:49:11.508815 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwfcs" event={"ID":"21d233b3-8a4c-4f3d-b89f-40af9b5ca381","Type":"ContainerStarted","Data":"58ee703e670bd512e5339d1a7dbe758607034fa5ecd610df69e4609aced1e2a3"} Oct 04 08:49:14 crc kubenswrapper[4969]: I1004 08:49:14.508026 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:14 crc kubenswrapper[4969]: I1004 08:49:14.508480 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:14 crc kubenswrapper[4969]: I1004 08:49:14.559710 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:14 crc kubenswrapper[4969]: I1004 08:49:14.583169 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nwfcs" podStartSLOduration=6.755775398 podStartE2EDuration="10.583154627s" podCreationTimestamp="2025-10-04 08:49:04 +0000 UTC" firstStartedPulling="2025-10-04 08:49:06.464656758 +0000 UTC m=+1974.218925582" lastFinishedPulling="2025-10-04 08:49:10.292035977 +0000 UTC m=+1978.046304811" observedRunningTime="2025-10-04 08:49:11.534334202 +0000 UTC m=+1979.288603056" watchObservedRunningTime="2025-10-04 08:49:14.583154627 +0000 UTC m=+1982.337423441" Oct 04 08:49:24 crc kubenswrapper[4969]: I1004 08:49:24.570346 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:24 crc kubenswrapper[4969]: I1004 08:49:24.643906 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nwfcs"] Oct 04 08:49:24 crc kubenswrapper[4969]: I1004 08:49:24.644167 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nwfcs" podUID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerName="registry-server" containerID="cri-o://58ee703e670bd512e5339d1a7dbe758607034fa5ecd610df69e4609aced1e2a3" gracePeriod=2 Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.634714 4969 generic.go:334] "Generic (PLEG): container finished" podID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerID="58ee703e670bd512e5339d1a7dbe758607034fa5ecd610df69e4609aced1e2a3" exitCode=0 Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.634768 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwfcs" event={"ID":"21d233b3-8a4c-4f3d-b89f-40af9b5ca381","Type":"ContainerDied","Data":"58ee703e670bd512e5339d1a7dbe758607034fa5ecd610df69e4609aced1e2a3"} Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.635389 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwfcs" event={"ID":"21d233b3-8a4c-4f3d-b89f-40af9b5ca381","Type":"ContainerDied","Data":"8ec7a8b87adb030ddb0115ecdbf50b933ba45a301a567a92b21db46ac187d181"} Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.635403 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ec7a8b87adb030ddb0115ecdbf50b933ba45a301a567a92b21db46ac187d181" Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.698089 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.725487 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-catalog-content\") pod \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.725595 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-utilities\") pod \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.725775 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdmsc\" (UniqueName: \"kubernetes.io/projected/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-kube-api-access-cdmsc\") pod \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\" (UID: \"21d233b3-8a4c-4f3d-b89f-40af9b5ca381\") " Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.726335 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-utilities" (OuterVolumeSpecName: "utilities") pod "21d233b3-8a4c-4f3d-b89f-40af9b5ca381" (UID: "21d233b3-8a4c-4f3d-b89f-40af9b5ca381"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.726843 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.733758 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-kube-api-access-cdmsc" (OuterVolumeSpecName: "kube-api-access-cdmsc") pod "21d233b3-8a4c-4f3d-b89f-40af9b5ca381" (UID: "21d233b3-8a4c-4f3d-b89f-40af9b5ca381"). InnerVolumeSpecName "kube-api-access-cdmsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.767159 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21d233b3-8a4c-4f3d-b89f-40af9b5ca381" (UID: "21d233b3-8a4c-4f3d-b89f-40af9b5ca381"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.829124 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdmsc\" (UniqueName: \"kubernetes.io/projected/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-kube-api-access-cdmsc\") on node \"crc\" DevicePath \"\"" Oct 04 08:49:25 crc kubenswrapper[4969]: I1004 08:49:25.829151 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d233b3-8a4c-4f3d-b89f-40af9b5ca381-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:49:26 crc kubenswrapper[4969]: I1004 08:49:26.644147 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nwfcs" Oct 04 08:49:26 crc kubenswrapper[4969]: I1004 08:49:26.684768 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nwfcs"] Oct 04 08:49:26 crc kubenswrapper[4969]: I1004 08:49:26.693729 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nwfcs"] Oct 04 08:49:27 crc kubenswrapper[4969]: I1004 08:49:27.069598 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" path="/var/lib/kubelet/pods/21d233b3-8a4c-4f3d-b89f-40af9b5ca381/volumes" Oct 04 08:49:42 crc kubenswrapper[4969]: I1004 08:49:42.814825 4969 generic.go:334] "Generic (PLEG): container finished" podID="19b4ae96-fa76-4d96-b4e0-d0058345fc23" containerID="79bee0fd376c80a3f61b36e0d2cf469dc7b5771fb1fe716cf332b0acc3a31478" exitCode=0 Oct 04 08:49:42 crc kubenswrapper[4969]: I1004 08:49:42.814929 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" event={"ID":"19b4ae96-fa76-4d96-b4e0-d0058345fc23","Type":"ContainerDied","Data":"79bee0fd376c80a3f61b36e0d2cf469dc7b5771fb1fe716cf332b0acc3a31478"} Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.381951 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.443672 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-ssh-key\") pod \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.443759 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbmfk\" (UniqueName: \"kubernetes.io/projected/19b4ae96-fa76-4d96-b4e0-d0058345fc23-kube-api-access-kbmfk\") pod \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.443783 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-inventory\") pod \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\" (UID: \"19b4ae96-fa76-4d96-b4e0-d0058345fc23\") " Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.451698 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19b4ae96-fa76-4d96-b4e0-d0058345fc23-kube-api-access-kbmfk" (OuterVolumeSpecName: "kube-api-access-kbmfk") pod "19b4ae96-fa76-4d96-b4e0-d0058345fc23" (UID: "19b4ae96-fa76-4d96-b4e0-d0058345fc23"). InnerVolumeSpecName "kube-api-access-kbmfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.476541 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-inventory" (OuterVolumeSpecName: "inventory") pod "19b4ae96-fa76-4d96-b4e0-d0058345fc23" (UID: "19b4ae96-fa76-4d96-b4e0-d0058345fc23"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.479863 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "19b4ae96-fa76-4d96-b4e0-d0058345fc23" (UID: "19b4ae96-fa76-4d96-b4e0-d0058345fc23"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.546071 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.546107 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbmfk\" (UniqueName: \"kubernetes.io/projected/19b4ae96-fa76-4d96-b4e0-d0058345fc23-kube-api-access-kbmfk\") on node \"crc\" DevicePath \"\"" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.546118 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b4ae96-fa76-4d96-b4e0-d0058345fc23-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.841823 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" event={"ID":"19b4ae96-fa76-4d96-b4e0-d0058345fc23","Type":"ContainerDied","Data":"8827fe252d97834df4c7d81a68be03118847cdfd78c4ea45aa4215c1911d6bf1"} Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.841875 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8827fe252d97834df4c7d81a68be03118847cdfd78c4ea45aa4215c1911d6bf1" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.841880 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t796n" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.945784 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dhdzp"] Oct 04 08:49:44 crc kubenswrapper[4969]: E1004 08:49:44.946446 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerName="extract-content" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.946558 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerName="extract-content" Oct 04 08:49:44 crc kubenswrapper[4969]: E1004 08:49:44.946630 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b4ae96-fa76-4d96-b4e0-d0058345fc23" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.947971 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b4ae96-fa76-4d96-b4e0-d0058345fc23" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:49:44 crc kubenswrapper[4969]: E1004 08:49:44.948051 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerName="extract-utilities" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.948121 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerName="extract-utilities" Oct 04 08:49:44 crc kubenswrapper[4969]: E1004 08:49:44.948215 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerName="registry-server" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.948297 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerName="registry-server" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.948665 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="21d233b3-8a4c-4f3d-b89f-40af9b5ca381" containerName="registry-server" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.948819 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b4ae96-fa76-4d96-b4e0-d0058345fc23" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.950158 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.956716 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.956715 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.956851 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.956859 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:49:44 crc kubenswrapper[4969]: I1004 08:49:44.967606 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dhdzp"] Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.055912 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dhdzp\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.056036 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dhdzp\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.056234 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbtdx\" (UniqueName: \"kubernetes.io/projected/b0fface3-5e9a-4a86-9863-8c7ad3c21151-kube-api-access-gbtdx\") pod \"ssh-known-hosts-edpm-deployment-dhdzp\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.157768 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbtdx\" (UniqueName: \"kubernetes.io/projected/b0fface3-5e9a-4a86-9863-8c7ad3c21151-kube-api-access-gbtdx\") pod \"ssh-known-hosts-edpm-deployment-dhdzp\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.157877 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dhdzp\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.157959 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dhdzp\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.163104 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dhdzp\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.164664 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dhdzp\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.173511 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbtdx\" (UniqueName: \"kubernetes.io/projected/b0fface3-5e9a-4a86-9863-8c7ad3c21151-kube-api-access-gbtdx\") pod \"ssh-known-hosts-edpm-deployment-dhdzp\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.274372 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.813755 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dhdzp"] Oct 04 08:49:45 crc kubenswrapper[4969]: I1004 08:49:45.855159 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" event={"ID":"b0fface3-5e9a-4a86-9863-8c7ad3c21151","Type":"ContainerStarted","Data":"47de70c2833af2662d615b3806f5452fa02a2b927926dcac6cdd8ee7b6e6735e"} Oct 04 08:49:46 crc kubenswrapper[4969]: I1004 08:49:46.864495 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" event={"ID":"b0fface3-5e9a-4a86-9863-8c7ad3c21151","Type":"ContainerStarted","Data":"92a08146ddf7fd169b4ec84d5a6a1ad8d87d36f58ed9725c5d5b610836e49d7a"} Oct 04 08:49:46 crc kubenswrapper[4969]: I1004 08:49:46.883381 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" podStartSLOduration=2.333288951 podStartE2EDuration="2.883359781s" podCreationTimestamp="2025-10-04 08:49:44 +0000 UTC" firstStartedPulling="2025-10-04 08:49:45.823506506 +0000 UTC m=+2013.577775320" lastFinishedPulling="2025-10-04 08:49:46.373577336 +0000 UTC m=+2014.127846150" observedRunningTime="2025-10-04 08:49:46.876686729 +0000 UTC m=+2014.630955553" watchObservedRunningTime="2025-10-04 08:49:46.883359781 +0000 UTC m=+2014.637628595" Oct 04 08:49:54 crc kubenswrapper[4969]: I1004 08:49:54.943412 4969 generic.go:334] "Generic (PLEG): container finished" podID="b0fface3-5e9a-4a86-9863-8c7ad3c21151" containerID="92a08146ddf7fd169b4ec84d5a6a1ad8d87d36f58ed9725c5d5b610836e49d7a" exitCode=0 Oct 04 08:49:54 crc kubenswrapper[4969]: I1004 08:49:54.943586 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" event={"ID":"b0fface3-5e9a-4a86-9863-8c7ad3c21151","Type":"ContainerDied","Data":"92a08146ddf7fd169b4ec84d5a6a1ad8d87d36f58ed9725c5d5b610836e49d7a"} Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.389743 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.498332 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbtdx\" (UniqueName: \"kubernetes.io/projected/b0fface3-5e9a-4a86-9863-8c7ad3c21151-kube-api-access-gbtdx\") pod \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.498522 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-ssh-key-openstack-edpm-ipam\") pod \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.498734 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-inventory-0\") pod \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\" (UID: \"b0fface3-5e9a-4a86-9863-8c7ad3c21151\") " Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.505156 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0fface3-5e9a-4a86-9863-8c7ad3c21151-kube-api-access-gbtdx" (OuterVolumeSpecName: "kube-api-access-gbtdx") pod "b0fface3-5e9a-4a86-9863-8c7ad3c21151" (UID: "b0fface3-5e9a-4a86-9863-8c7ad3c21151"). InnerVolumeSpecName "kube-api-access-gbtdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.531637 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b0fface3-5e9a-4a86-9863-8c7ad3c21151" (UID: "b0fface3-5e9a-4a86-9863-8c7ad3c21151"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.543293 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b0fface3-5e9a-4a86-9863-8c7ad3c21151" (UID: "b0fface3-5e9a-4a86-9863-8c7ad3c21151"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.602288 4969 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.602331 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbtdx\" (UniqueName: \"kubernetes.io/projected/b0fface3-5e9a-4a86-9863-8c7ad3c21151-kube-api-access-gbtdx\") on node \"crc\" DevicePath \"\"" Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.602348 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0fface3-5e9a-4a86-9863-8c7ad3c21151-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.971719 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" event={"ID":"b0fface3-5e9a-4a86-9863-8c7ad3c21151","Type":"ContainerDied","Data":"47de70c2833af2662d615b3806f5452fa02a2b927926dcac6cdd8ee7b6e6735e"} Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.971760 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47de70c2833af2662d615b3806f5452fa02a2b927926dcac6cdd8ee7b6e6735e" Oct 04 08:49:56 crc kubenswrapper[4969]: I1004 08:49:56.971759 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dhdzp" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.050070 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg"] Oct 04 08:49:57 crc kubenswrapper[4969]: E1004 08:49:57.050680 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fface3-5e9a-4a86-9863-8c7ad3c21151" containerName="ssh-known-hosts-edpm-deployment" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.050746 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fface3-5e9a-4a86-9863-8c7ad3c21151" containerName="ssh-known-hosts-edpm-deployment" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.051025 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0fface3-5e9a-4a86-9863-8c7ad3c21151" containerName="ssh-known-hosts-edpm-deployment" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.051823 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.054123 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.054385 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.054480 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.055410 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.070729 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg"] Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.213660 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkhx5\" (UniqueName: \"kubernetes.io/projected/57aac488-900c-47d7-abc5-0349f1004655-kube-api-access-lkhx5\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kclpg\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.213811 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kclpg\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.213978 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kclpg\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.315940 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kclpg\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.316024 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkhx5\" (UniqueName: \"kubernetes.io/projected/57aac488-900c-47d7-abc5-0349f1004655-kube-api-access-lkhx5\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kclpg\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.316086 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kclpg\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.321713 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kclpg\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.331080 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kclpg\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.340556 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkhx5\" (UniqueName: \"kubernetes.io/projected/57aac488-900c-47d7-abc5-0349f1004655-kube-api-access-lkhx5\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kclpg\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.371078 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.954650 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg"] Oct 04 08:49:57 crc kubenswrapper[4969]: I1004 08:49:57.983504 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" event={"ID":"57aac488-900c-47d7-abc5-0349f1004655","Type":"ContainerStarted","Data":"af5f7726cbd45780c3b0d3df945e2ea924f6965d03662b610df4ed065ff9a7bd"} Oct 04 08:49:59 crc kubenswrapper[4969]: I1004 08:49:59.000246 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" event={"ID":"57aac488-900c-47d7-abc5-0349f1004655","Type":"ContainerStarted","Data":"9057f709da35911c7b613f9e786dda154d990aa659cab7e988c8dc1b442514cf"} Oct 04 08:49:59 crc kubenswrapper[4969]: I1004 08:49:59.032194 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" podStartSLOduration=1.622606078 podStartE2EDuration="2.032167792s" podCreationTimestamp="2025-10-04 08:49:57 +0000 UTC" firstStartedPulling="2025-10-04 08:49:57.955815285 +0000 UTC m=+2025.710084099" lastFinishedPulling="2025-10-04 08:49:58.365376999 +0000 UTC m=+2026.119645813" observedRunningTime="2025-10-04 08:49:59.021612161 +0000 UTC m=+2026.775881045" watchObservedRunningTime="2025-10-04 08:49:59.032167792 +0000 UTC m=+2026.786436636" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.366088 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dkd6p"] Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.370078 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.395532 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dkd6p"] Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.507943 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-utilities\") pod \"community-operators-dkd6p\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.508044 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs29h\" (UniqueName: \"kubernetes.io/projected/4c111897-4e30-4003-ae4a-51862de2ac51-kube-api-access-xs29h\") pod \"community-operators-dkd6p\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.508121 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-catalog-content\") pod \"community-operators-dkd6p\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.610248 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-catalog-content\") pod \"community-operators-dkd6p\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.610394 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-utilities\") pod \"community-operators-dkd6p\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.610449 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs29h\" (UniqueName: \"kubernetes.io/projected/4c111897-4e30-4003-ae4a-51862de2ac51-kube-api-access-xs29h\") pod \"community-operators-dkd6p\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.610976 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-catalog-content\") pod \"community-operators-dkd6p\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.612168 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-utilities\") pod \"community-operators-dkd6p\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.642372 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs29h\" (UniqueName: \"kubernetes.io/projected/4c111897-4e30-4003-ae4a-51862de2ac51-kube-api-access-xs29h\") pod \"community-operators-dkd6p\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:06 crc kubenswrapper[4969]: I1004 08:50:06.703647 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:07 crc kubenswrapper[4969]: I1004 08:50:07.273127 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dkd6p"] Oct 04 08:50:07 crc kubenswrapper[4969]: W1004 08:50:07.275858 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c111897_4e30_4003_ae4a_51862de2ac51.slice/crio-fc58fa6ed12cd6a5e05bf2cd2f4ecb25821e1cad8e1817e1babe6fe274a557fc WatchSource:0}: Error finding container fc58fa6ed12cd6a5e05bf2cd2f4ecb25821e1cad8e1817e1babe6fe274a557fc: Status 404 returned error can't find the container with id fc58fa6ed12cd6a5e05bf2cd2f4ecb25821e1cad8e1817e1babe6fe274a557fc Oct 04 08:50:08 crc kubenswrapper[4969]: I1004 08:50:08.091766 4969 generic.go:334] "Generic (PLEG): container finished" podID="4c111897-4e30-4003-ae4a-51862de2ac51" containerID="eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f" exitCode=0 Oct 04 08:50:08 crc kubenswrapper[4969]: I1004 08:50:08.091852 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkd6p" event={"ID":"4c111897-4e30-4003-ae4a-51862de2ac51","Type":"ContainerDied","Data":"eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f"} Oct 04 08:50:08 crc kubenswrapper[4969]: I1004 08:50:08.092009 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkd6p" event={"ID":"4c111897-4e30-4003-ae4a-51862de2ac51","Type":"ContainerStarted","Data":"fc58fa6ed12cd6a5e05bf2cd2f4ecb25821e1cad8e1817e1babe6fe274a557fc"} Oct 04 08:50:09 crc kubenswrapper[4969]: I1004 08:50:09.107936 4969 generic.go:334] "Generic (PLEG): container finished" podID="57aac488-900c-47d7-abc5-0349f1004655" containerID="9057f709da35911c7b613f9e786dda154d990aa659cab7e988c8dc1b442514cf" exitCode=0 Oct 04 08:50:09 crc kubenswrapper[4969]: I1004 08:50:09.108042 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" event={"ID":"57aac488-900c-47d7-abc5-0349f1004655","Type":"ContainerDied","Data":"9057f709da35911c7b613f9e786dda154d990aa659cab7e988c8dc1b442514cf"} Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.121596 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkd6p" event={"ID":"4c111897-4e30-4003-ae4a-51862de2ac51","Type":"ContainerStarted","Data":"795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987"} Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.553499 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.695034 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-ssh-key\") pod \"57aac488-900c-47d7-abc5-0349f1004655\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.695123 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-inventory\") pod \"57aac488-900c-47d7-abc5-0349f1004655\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.695218 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkhx5\" (UniqueName: \"kubernetes.io/projected/57aac488-900c-47d7-abc5-0349f1004655-kube-api-access-lkhx5\") pod \"57aac488-900c-47d7-abc5-0349f1004655\" (UID: \"57aac488-900c-47d7-abc5-0349f1004655\") " Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.705691 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57aac488-900c-47d7-abc5-0349f1004655-kube-api-access-lkhx5" (OuterVolumeSpecName: "kube-api-access-lkhx5") pod "57aac488-900c-47d7-abc5-0349f1004655" (UID: "57aac488-900c-47d7-abc5-0349f1004655"). InnerVolumeSpecName "kube-api-access-lkhx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.729949 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-inventory" (OuterVolumeSpecName: "inventory") pod "57aac488-900c-47d7-abc5-0349f1004655" (UID: "57aac488-900c-47d7-abc5-0349f1004655"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.734026 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "57aac488-900c-47d7-abc5-0349f1004655" (UID: "57aac488-900c-47d7-abc5-0349f1004655"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.797917 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.797965 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkhx5\" (UniqueName: \"kubernetes.io/projected/57aac488-900c-47d7-abc5-0349f1004655-kube-api-access-lkhx5\") on node \"crc\" DevicePath \"\"" Oct 04 08:50:10 crc kubenswrapper[4969]: I1004 08:50:10.797981 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57aac488-900c-47d7-abc5-0349f1004655-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.139516 4969 generic.go:334] "Generic (PLEG): container finished" podID="4c111897-4e30-4003-ae4a-51862de2ac51" containerID="795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987" exitCode=0 Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.139619 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkd6p" event={"ID":"4c111897-4e30-4003-ae4a-51862de2ac51","Type":"ContainerDied","Data":"795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987"} Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.144358 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" event={"ID":"57aac488-900c-47d7-abc5-0349f1004655","Type":"ContainerDied","Data":"af5f7726cbd45780c3b0d3df945e2ea924f6965d03662b610df4ed065ff9a7bd"} Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.144390 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af5f7726cbd45780c3b0d3df945e2ea924f6965d03662b610df4ed065ff9a7bd" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.144537 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kclpg" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.219896 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm"] Oct 04 08:50:11 crc kubenswrapper[4969]: E1004 08:50:11.220356 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57aac488-900c-47d7-abc5-0349f1004655" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.220375 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="57aac488-900c-47d7-abc5-0349f1004655" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.220611 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="57aac488-900c-47d7-abc5-0349f1004655" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.221467 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.223350 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.223653 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.223719 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.224143 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.234972 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm"] Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.410797 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.411232 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.411332 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8wpv\" (UniqueName: \"kubernetes.io/projected/942f1580-e8dc-4f73-aa37-e5a2cde068c6-kube-api-access-l8wpv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.513606 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.513686 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8wpv\" (UniqueName: \"kubernetes.io/projected/942f1580-e8dc-4f73-aa37-e5a2cde068c6-kube-api-access-l8wpv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.513787 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.520678 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.528615 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.532344 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8wpv\" (UniqueName: \"kubernetes.io/projected/942f1580-e8dc-4f73-aa37-e5a2cde068c6-kube-api-access-l8wpv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:11 crc kubenswrapper[4969]: I1004 08:50:11.543170 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:12 crc kubenswrapper[4969]: I1004 08:50:12.183548 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm"] Oct 04 08:50:13 crc kubenswrapper[4969]: I1004 08:50:13.166792 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkd6p" event={"ID":"4c111897-4e30-4003-ae4a-51862de2ac51","Type":"ContainerStarted","Data":"34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4"} Oct 04 08:50:13 crc kubenswrapper[4969]: I1004 08:50:13.168757 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" event={"ID":"942f1580-e8dc-4f73-aa37-e5a2cde068c6","Type":"ContainerStarted","Data":"36d185399c380a52edef987be50c538a42be0850ea35d3bf8c62b3bae016ebf7"} Oct 04 08:50:13 crc kubenswrapper[4969]: I1004 08:50:13.192179 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dkd6p" podStartSLOduration=3.264313241 podStartE2EDuration="7.192150044s" podCreationTimestamp="2025-10-04 08:50:06 +0000 UTC" firstStartedPulling="2025-10-04 08:50:08.095350719 +0000 UTC m=+2035.849619543" lastFinishedPulling="2025-10-04 08:50:12.023187512 +0000 UTC m=+2039.777456346" observedRunningTime="2025-10-04 08:50:13.18536558 +0000 UTC m=+2040.939634654" watchObservedRunningTime="2025-10-04 08:50:13.192150044 +0000 UTC m=+2040.946418898" Oct 04 08:50:13 crc kubenswrapper[4969]: I1004 08:50:13.517887 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:50:14 crc kubenswrapper[4969]: I1004 08:50:14.178534 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" event={"ID":"942f1580-e8dc-4f73-aa37-e5a2cde068c6","Type":"ContainerStarted","Data":"3ad0e222497276ab0e6e458ab36988ab20db6681bdeacadf2f042c8cab4690f8"} Oct 04 08:50:14 crc kubenswrapper[4969]: I1004 08:50:14.194589 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" podStartSLOduration=1.8659594579999998 podStartE2EDuration="3.194563055s" podCreationTimestamp="2025-10-04 08:50:11 +0000 UTC" firstStartedPulling="2025-10-04 08:50:12.186163492 +0000 UTC m=+2039.940432306" lastFinishedPulling="2025-10-04 08:50:13.514767089 +0000 UTC m=+2041.269035903" observedRunningTime="2025-10-04 08:50:14.19320766 +0000 UTC m=+2041.947476484" watchObservedRunningTime="2025-10-04 08:50:14.194563055 +0000 UTC m=+2041.948831909" Oct 04 08:50:16 crc kubenswrapper[4969]: I1004 08:50:16.704259 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:16 crc kubenswrapper[4969]: I1004 08:50:16.704736 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:16 crc kubenswrapper[4969]: I1004 08:50:16.794006 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:17 crc kubenswrapper[4969]: I1004 08:50:17.289928 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:17 crc kubenswrapper[4969]: I1004 08:50:17.353585 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dkd6p"] Oct 04 08:50:19 crc kubenswrapper[4969]: I1004 08:50:19.226170 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dkd6p" podUID="4c111897-4e30-4003-ae4a-51862de2ac51" containerName="registry-server" containerID="cri-o://34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4" gracePeriod=2 Oct 04 08:50:19 crc kubenswrapper[4969]: I1004 08:50:19.667068 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:50:19 crc kubenswrapper[4969]: I1004 08:50:19.667555 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.233627 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.235876 4969 generic.go:334] "Generic (PLEG): container finished" podID="4c111897-4e30-4003-ae4a-51862de2ac51" containerID="34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4" exitCode=0 Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.235911 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkd6p" event={"ID":"4c111897-4e30-4003-ae4a-51862de2ac51","Type":"ContainerDied","Data":"34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4"} Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.235938 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkd6p" event={"ID":"4c111897-4e30-4003-ae4a-51862de2ac51","Type":"ContainerDied","Data":"fc58fa6ed12cd6a5e05bf2cd2f4ecb25821e1cad8e1817e1babe6fe274a557fc"} Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.235978 4969 scope.go:117] "RemoveContainer" containerID="34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.256665 4969 scope.go:117] "RemoveContainer" containerID="795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.289761 4969 scope.go:117] "RemoveContainer" containerID="eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.327846 4969 scope.go:117] "RemoveContainer" containerID="34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4" Oct 04 08:50:20 crc kubenswrapper[4969]: E1004 08:50:20.328253 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4\": container with ID starting with 34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4 not found: ID does not exist" containerID="34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.328286 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4"} err="failed to get container status \"34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4\": rpc error: code = NotFound desc = could not find container \"34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4\": container with ID starting with 34e0a9f0652af354e8f48bcff5efcb06280d6ba1cbb5ef5d25a3859f3933a5e4 not found: ID does not exist" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.328306 4969 scope.go:117] "RemoveContainer" containerID="795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987" Oct 04 08:50:20 crc kubenswrapper[4969]: E1004 08:50:20.328660 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987\": container with ID starting with 795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987 not found: ID does not exist" containerID="795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.328708 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987"} err="failed to get container status \"795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987\": rpc error: code = NotFound desc = could not find container \"795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987\": container with ID starting with 795cb4a234fb9e6088293d83b2ae330ff78c2bab7af694538d8f7f095cd8a987 not found: ID does not exist" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.328747 4969 scope.go:117] "RemoveContainer" containerID="eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f" Oct 04 08:50:20 crc kubenswrapper[4969]: E1004 08:50:20.329100 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f\": container with ID starting with eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f not found: ID does not exist" containerID="eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.329128 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f"} err="failed to get container status \"eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f\": rpc error: code = NotFound desc = could not find container \"eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f\": container with ID starting with eb83549eee622f079512e6b880f15a81bee299a26145e8e137601de992bf531f not found: ID does not exist" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.424942 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-utilities\") pod \"4c111897-4e30-4003-ae4a-51862de2ac51\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.425014 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs29h\" (UniqueName: \"kubernetes.io/projected/4c111897-4e30-4003-ae4a-51862de2ac51-kube-api-access-xs29h\") pod \"4c111897-4e30-4003-ae4a-51862de2ac51\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.425122 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-catalog-content\") pod \"4c111897-4e30-4003-ae4a-51862de2ac51\" (UID: \"4c111897-4e30-4003-ae4a-51862de2ac51\") " Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.426001 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-utilities" (OuterVolumeSpecName: "utilities") pod "4c111897-4e30-4003-ae4a-51862de2ac51" (UID: "4c111897-4e30-4003-ae4a-51862de2ac51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.436328 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c111897-4e30-4003-ae4a-51862de2ac51-kube-api-access-xs29h" (OuterVolumeSpecName: "kube-api-access-xs29h") pod "4c111897-4e30-4003-ae4a-51862de2ac51" (UID: "4c111897-4e30-4003-ae4a-51862de2ac51"). InnerVolumeSpecName "kube-api-access-xs29h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.528703 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:50:20 crc kubenswrapper[4969]: I1004 08:50:20.528751 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs29h\" (UniqueName: \"kubernetes.io/projected/4c111897-4e30-4003-ae4a-51862de2ac51-kube-api-access-xs29h\") on node \"crc\" DevicePath \"\"" Oct 04 08:50:21 crc kubenswrapper[4969]: I1004 08:50:21.249688 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkd6p" Oct 04 08:50:23 crc kubenswrapper[4969]: I1004 08:50:23.987490 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c111897-4e30-4003-ae4a-51862de2ac51" (UID: "4c111897-4e30-4003-ae4a-51862de2ac51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:50:24 crc kubenswrapper[4969]: I1004 08:50:24.007378 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c111897-4e30-4003-ae4a-51862de2ac51-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:50:24 crc kubenswrapper[4969]: I1004 08:50:24.296385 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dkd6p"] Oct 04 08:50:24 crc kubenswrapper[4969]: I1004 08:50:24.307248 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dkd6p"] Oct 04 08:50:25 crc kubenswrapper[4969]: I1004 08:50:25.087647 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c111897-4e30-4003-ae4a-51862de2ac51" path="/var/lib/kubelet/pods/4c111897-4e30-4003-ae4a-51862de2ac51/volumes" Oct 04 08:50:25 crc kubenswrapper[4969]: I1004 08:50:25.283147 4969 generic.go:334] "Generic (PLEG): container finished" podID="942f1580-e8dc-4f73-aa37-e5a2cde068c6" containerID="3ad0e222497276ab0e6e458ab36988ab20db6681bdeacadf2f042c8cab4690f8" exitCode=0 Oct 04 08:50:25 crc kubenswrapper[4969]: I1004 08:50:25.283207 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" event={"ID":"942f1580-e8dc-4f73-aa37-e5a2cde068c6","Type":"ContainerDied","Data":"3ad0e222497276ab0e6e458ab36988ab20db6681bdeacadf2f042c8cab4690f8"} Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.747222 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.863691 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-ssh-key\") pod \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.863831 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-inventory\") pod \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.863921 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8wpv\" (UniqueName: \"kubernetes.io/projected/942f1580-e8dc-4f73-aa37-e5a2cde068c6-kube-api-access-l8wpv\") pod \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\" (UID: \"942f1580-e8dc-4f73-aa37-e5a2cde068c6\") " Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.870703 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/942f1580-e8dc-4f73-aa37-e5a2cde068c6-kube-api-access-l8wpv" (OuterVolumeSpecName: "kube-api-access-l8wpv") pod "942f1580-e8dc-4f73-aa37-e5a2cde068c6" (UID: "942f1580-e8dc-4f73-aa37-e5a2cde068c6"). InnerVolumeSpecName "kube-api-access-l8wpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.893378 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-inventory" (OuterVolumeSpecName: "inventory") pod "942f1580-e8dc-4f73-aa37-e5a2cde068c6" (UID: "942f1580-e8dc-4f73-aa37-e5a2cde068c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.897599 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "942f1580-e8dc-4f73-aa37-e5a2cde068c6" (UID: "942f1580-e8dc-4f73-aa37-e5a2cde068c6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.965678 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8wpv\" (UniqueName: \"kubernetes.io/projected/942f1580-e8dc-4f73-aa37-e5a2cde068c6-kube-api-access-l8wpv\") on node \"crc\" DevicePath \"\"" Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.965708 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:50:26 crc kubenswrapper[4969]: I1004 08:50:26.965720 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/942f1580-e8dc-4f73-aa37-e5a2cde068c6-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.313176 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" event={"ID":"942f1580-e8dc-4f73-aa37-e5a2cde068c6","Type":"ContainerDied","Data":"36d185399c380a52edef987be50c538a42be0850ea35d3bf8c62b3bae016ebf7"} Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.313573 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36d185399c380a52edef987be50c538a42be0850ea35d3bf8c62b3bae016ebf7" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.313265 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.421946 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw"] Oct 04 08:50:27 crc kubenswrapper[4969]: E1004 08:50:27.422504 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c111897-4e30-4003-ae4a-51862de2ac51" containerName="extract-content" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.422527 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c111897-4e30-4003-ae4a-51862de2ac51" containerName="extract-content" Oct 04 08:50:27 crc kubenswrapper[4969]: E1004 08:50:27.422554 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c111897-4e30-4003-ae4a-51862de2ac51" containerName="registry-server" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.422562 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c111897-4e30-4003-ae4a-51862de2ac51" containerName="registry-server" Oct 04 08:50:27 crc kubenswrapper[4969]: E1004 08:50:27.422586 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="942f1580-e8dc-4f73-aa37-e5a2cde068c6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.422595 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="942f1580-e8dc-4f73-aa37-e5a2cde068c6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:50:27 crc kubenswrapper[4969]: E1004 08:50:27.422607 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c111897-4e30-4003-ae4a-51862de2ac51" containerName="extract-utilities" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.422615 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c111897-4e30-4003-ae4a-51862de2ac51" containerName="extract-utilities" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.422846 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="942f1580-e8dc-4f73-aa37-e5a2cde068c6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.422897 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c111897-4e30-4003-ae4a-51862de2ac51" containerName="registry-server" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.425041 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.433123 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw"] Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.434203 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.434594 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.435113 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.435352 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.437754 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.438156 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.438954 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.439117 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.478913 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.479166 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.479361 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.479498 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.479579 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.479664 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.479755 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kpsn\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-kube-api-access-5kpsn\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.479849 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.479936 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.480026 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.480140 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.480270 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.480407 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.480545 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581344 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581465 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581519 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581578 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581646 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581680 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581707 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581733 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581758 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581790 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581818 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kpsn\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-kube-api-access-5kpsn\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581857 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581894 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.581918 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.586160 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.586667 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.586677 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.587460 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.587824 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.588221 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.588844 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.590037 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.596229 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.597010 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.600371 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.601752 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.601998 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.606619 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kpsn\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-kube-api-access-5kpsn\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:27 crc kubenswrapper[4969]: I1004 08:50:27.758991 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:50:28 crc kubenswrapper[4969]: I1004 08:50:28.393778 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw"] Oct 04 08:50:28 crc kubenswrapper[4969]: I1004 08:50:28.402754 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 08:50:29 crc kubenswrapper[4969]: I1004 08:50:29.331243 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" event={"ID":"a7f1a5ed-9324-4eb1-b561-d65fca042f4d","Type":"ContainerStarted","Data":"4a51b9412bc23885b38bc334616487b44d5a96f3d81db66e904da54ec5ee3ba6"} Oct 04 08:50:33 crc kubenswrapper[4969]: I1004 08:50:33.379475 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" event={"ID":"a7f1a5ed-9324-4eb1-b561-d65fca042f4d","Type":"ContainerStarted","Data":"8cdb5aa04cf4bc3e6daf3529d6b73a0b61ebb7f56e8334bfdb2ddab3734a99a2"} Oct 04 08:50:33 crc kubenswrapper[4969]: I1004 08:50:33.409498 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" podStartSLOduration=1.940354593 podStartE2EDuration="6.4094802s" podCreationTimestamp="2025-10-04 08:50:27 +0000 UTC" firstStartedPulling="2025-10-04 08:50:28.402478868 +0000 UTC m=+2056.156747692" lastFinishedPulling="2025-10-04 08:50:32.871604445 +0000 UTC m=+2060.625873299" observedRunningTime="2025-10-04 08:50:33.404515154 +0000 UTC m=+2061.158783988" watchObservedRunningTime="2025-10-04 08:50:33.4094802 +0000 UTC m=+2061.163749024" Oct 04 08:50:49 crc kubenswrapper[4969]: I1004 08:50:49.670038 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:50:49 crc kubenswrapper[4969]: I1004 08:50:49.670661 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.443134 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-45pwx"] Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.446566 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.453202 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-45pwx"] Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.491599 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-utilities\") pod \"redhat-operators-45pwx\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.491688 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9zt6\" (UniqueName: \"kubernetes.io/projected/a4a9a67e-74d4-42bf-a563-555497bcd55c-kube-api-access-p9zt6\") pod \"redhat-operators-45pwx\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.491884 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-catalog-content\") pod \"redhat-operators-45pwx\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.596550 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-utilities\") pod \"redhat-operators-45pwx\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.596640 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9zt6\" (UniqueName: \"kubernetes.io/projected/a4a9a67e-74d4-42bf-a563-555497bcd55c-kube-api-access-p9zt6\") pod \"redhat-operators-45pwx\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.596683 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-catalog-content\") pod \"redhat-operators-45pwx\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.597155 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-catalog-content\") pod \"redhat-operators-45pwx\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.597441 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-utilities\") pod \"redhat-operators-45pwx\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.628281 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9zt6\" (UniqueName: \"kubernetes.io/projected/a4a9a67e-74d4-42bf-a563-555497bcd55c-kube-api-access-p9zt6\") pod \"redhat-operators-45pwx\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:10 crc kubenswrapper[4969]: I1004 08:51:10.770592 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:11 crc kubenswrapper[4969]: I1004 08:51:11.233697 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-45pwx"] Oct 04 08:51:11 crc kubenswrapper[4969]: I1004 08:51:11.739465 4969 generic.go:334] "Generic (PLEG): container finished" podID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerID="1e5f746da655d545ec86577e59af0e58a481fd9062f6b9057ca7f65efa29aa4b" exitCode=0 Oct 04 08:51:11 crc kubenswrapper[4969]: I1004 08:51:11.739539 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45pwx" event={"ID":"a4a9a67e-74d4-42bf-a563-555497bcd55c","Type":"ContainerDied","Data":"1e5f746da655d545ec86577e59af0e58a481fd9062f6b9057ca7f65efa29aa4b"} Oct 04 08:51:11 crc kubenswrapper[4969]: I1004 08:51:11.739820 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45pwx" event={"ID":"a4a9a67e-74d4-42bf-a563-555497bcd55c","Type":"ContainerStarted","Data":"0f21b36a50c4fad5facce2283608a1b9d8651dd2e3b4740985bed843f20c949a"} Oct 04 08:51:13 crc kubenswrapper[4969]: I1004 08:51:13.767382 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45pwx" event={"ID":"a4a9a67e-74d4-42bf-a563-555497bcd55c","Type":"ContainerStarted","Data":"8adb734b3cec2d57e79bc0958e996165b14ea64a109cfe0992db23a528451ece"} Oct 04 08:51:15 crc kubenswrapper[4969]: I1004 08:51:15.786078 4969 generic.go:334] "Generic (PLEG): container finished" podID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerID="8adb734b3cec2d57e79bc0958e996165b14ea64a109cfe0992db23a528451ece" exitCode=0 Oct 04 08:51:15 crc kubenswrapper[4969]: I1004 08:51:15.786169 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45pwx" event={"ID":"a4a9a67e-74d4-42bf-a563-555497bcd55c","Type":"ContainerDied","Data":"8adb734b3cec2d57e79bc0958e996165b14ea64a109cfe0992db23a528451ece"} Oct 04 08:51:17 crc kubenswrapper[4969]: I1004 08:51:17.814294 4969 generic.go:334] "Generic (PLEG): container finished" podID="a7f1a5ed-9324-4eb1-b561-d65fca042f4d" containerID="8cdb5aa04cf4bc3e6daf3529d6b73a0b61ebb7f56e8334bfdb2ddab3734a99a2" exitCode=0 Oct 04 08:51:17 crc kubenswrapper[4969]: I1004 08:51:17.814449 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" event={"ID":"a7f1a5ed-9324-4eb1-b561-d65fca042f4d","Type":"ContainerDied","Data":"8cdb5aa04cf4bc3e6daf3529d6b73a0b61ebb7f56e8334bfdb2ddab3734a99a2"} Oct 04 08:51:18 crc kubenswrapper[4969]: I1004 08:51:18.825241 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45pwx" event={"ID":"a4a9a67e-74d4-42bf-a563-555497bcd55c","Type":"ContainerStarted","Data":"c11814262f9039b86733dc8a0c4195c9efb64f7c044f622fa973be070ce5e568"} Oct 04 08:51:18 crc kubenswrapper[4969]: I1004 08:51:18.850420 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-45pwx" podStartSLOduration=3.121205759 podStartE2EDuration="8.850395095s" podCreationTimestamp="2025-10-04 08:51:10 +0000 UTC" firstStartedPulling="2025-10-04 08:51:11.743869092 +0000 UTC m=+2099.498137906" lastFinishedPulling="2025-10-04 08:51:17.473058388 +0000 UTC m=+2105.227327242" observedRunningTime="2025-10-04 08:51:18.843379785 +0000 UTC m=+2106.597648599" watchObservedRunningTime="2025-10-04 08:51:18.850395095 +0000 UTC m=+2106.604663909" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.247334 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.417961 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418035 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-nova-combined-ca-bundle\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418078 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418152 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ssh-key\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418198 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-libvirt-combined-ca-bundle\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418236 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418278 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418323 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-neutron-metadata-combined-ca-bundle\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418350 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-telemetry-combined-ca-bundle\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418404 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kpsn\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-kube-api-access-5kpsn\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418484 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-repo-setup-combined-ca-bundle\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418517 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-inventory\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418555 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ovn-combined-ca-bundle\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.418614 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-bootstrap-combined-ca-bundle\") pod \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\" (UID: \"a7f1a5ed-9324-4eb1-b561-d65fca042f4d\") " Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.424848 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.424889 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.424973 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.425161 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.425445 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.426632 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.427652 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.427994 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.428209 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.428802 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.431626 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.431729 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-kube-api-access-5kpsn" (OuterVolumeSpecName: "kube-api-access-5kpsn") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "kube-api-access-5kpsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.454267 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.455656 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-inventory" (OuterVolumeSpecName: "inventory") pod "a7f1a5ed-9324-4eb1-b561-d65fca042f4d" (UID: "a7f1a5ed-9324-4eb1-b561-d65fca042f4d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521101 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521134 4969 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521148 4969 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521158 4969 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521169 4969 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521178 4969 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521187 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kpsn\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-kube-api-access-5kpsn\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521196 4969 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521204 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521213 4969 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521222 4969 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521231 4969 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521239 4969 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.521251 4969 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a7f1a5ed-9324-4eb1-b561-d65fca042f4d-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.666919 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.667162 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.667200 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.667918 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f87372b3dbcc42c8848b2ff115cee7651dab7a8ca5bb7cad720f3bfdc85a0f10"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.667965 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://f87372b3dbcc42c8848b2ff115cee7651dab7a8ca5bb7cad720f3bfdc85a0f10" gracePeriod=600 Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.848777 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" event={"ID":"a7f1a5ed-9324-4eb1-b561-d65fca042f4d","Type":"ContainerDied","Data":"4a51b9412bc23885b38bc334616487b44d5a96f3d81db66e904da54ec5ee3ba6"} Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.848827 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a51b9412bc23885b38bc334616487b44d5a96f3d81db66e904da54ec5ee3ba6" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.848844 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw" Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.851289 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="f87372b3dbcc42c8848b2ff115cee7651dab7a8ca5bb7cad720f3bfdc85a0f10" exitCode=0 Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.851341 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"f87372b3dbcc42c8848b2ff115cee7651dab7a8ca5bb7cad720f3bfdc85a0f10"} Oct 04 08:51:19 crc kubenswrapper[4969]: I1004 08:51:19.851387 4969 scope.go:117] "RemoveContainer" containerID="297d2608678074f9c2b7d13b18eabd1645b17ea3ab1764209dc81d572cf9b8d6" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.044291 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl"] Oct 04 08:51:20 crc kubenswrapper[4969]: E1004 08:51:20.044710 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f1a5ed-9324-4eb1-b561-d65fca042f4d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.044726 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f1a5ed-9324-4eb1-b561-d65fca042f4d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.044916 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f1a5ed-9324-4eb1-b561-d65fca042f4d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.045611 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.048611 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.048908 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.049800 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.050757 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.062738 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.065932 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl"] Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.236156 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.236449 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.236500 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.236546 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.236680 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfx4k\" (UniqueName: \"kubernetes.io/projected/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-kube-api-access-bfx4k\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.338599 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.338711 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.338810 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.338887 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.338977 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfx4k\" (UniqueName: \"kubernetes.io/projected/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-kube-api-access-bfx4k\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.339844 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.345063 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.345214 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.350906 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.362967 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfx4k\" (UniqueName: \"kubernetes.io/projected/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-kube-api-access-bfx4k\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4hrwl\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.363804 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.770878 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.771831 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.862886 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916"} Oct 04 08:51:20 crc kubenswrapper[4969]: I1004 08:51:20.899466 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl"] Oct 04 08:51:21 crc kubenswrapper[4969]: I1004 08:51:21.829521 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-45pwx" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerName="registry-server" probeResult="failure" output=< Oct 04 08:51:21 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 08:51:21 crc kubenswrapper[4969]: > Oct 04 08:51:21 crc kubenswrapper[4969]: I1004 08:51:21.875992 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" event={"ID":"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e","Type":"ContainerStarted","Data":"3a6a2defa3245cfda56f2d45274087690d4753527d32ae05fc55931b2c1b3d29"} Oct 04 08:51:22 crc kubenswrapper[4969]: I1004 08:51:22.888788 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" event={"ID":"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e","Type":"ContainerStarted","Data":"a9fff868b811ac08167d559b3b3a21b7e22584d8860c0a4db72eff67465aae74"} Oct 04 08:51:22 crc kubenswrapper[4969]: I1004 08:51:22.913880 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" podStartSLOduration=1.481393107 podStartE2EDuration="2.913856387s" podCreationTimestamp="2025-10-04 08:51:20 +0000 UTC" firstStartedPulling="2025-10-04 08:51:20.918602972 +0000 UTC m=+2108.672871786" lastFinishedPulling="2025-10-04 08:51:22.351066222 +0000 UTC m=+2110.105335066" observedRunningTime="2025-10-04 08:51:22.90501285 +0000 UTC m=+2110.659281664" watchObservedRunningTime="2025-10-04 08:51:22.913856387 +0000 UTC m=+2110.668125191" Oct 04 08:51:30 crc kubenswrapper[4969]: I1004 08:51:30.826790 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:30 crc kubenswrapper[4969]: I1004 08:51:30.901402 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:31 crc kubenswrapper[4969]: I1004 08:51:31.102476 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-45pwx"] Oct 04 08:51:31 crc kubenswrapper[4969]: I1004 08:51:31.983649 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-45pwx" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerName="registry-server" containerID="cri-o://c11814262f9039b86733dc8a0c4195c9efb64f7c044f622fa973be070ce5e568" gracePeriod=2 Oct 04 08:51:32 crc kubenswrapper[4969]: I1004 08:51:32.995472 4969 generic.go:334] "Generic (PLEG): container finished" podID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerID="c11814262f9039b86733dc8a0c4195c9efb64f7c044f622fa973be070ce5e568" exitCode=0 Oct 04 08:51:32 crc kubenswrapper[4969]: I1004 08:51:32.995563 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45pwx" event={"ID":"a4a9a67e-74d4-42bf-a563-555497bcd55c","Type":"ContainerDied","Data":"c11814262f9039b86733dc8a0c4195c9efb64f7c044f622fa973be070ce5e568"} Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.214677 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.321959 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9zt6\" (UniqueName: \"kubernetes.io/projected/a4a9a67e-74d4-42bf-a563-555497bcd55c-kube-api-access-p9zt6\") pod \"a4a9a67e-74d4-42bf-a563-555497bcd55c\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.322186 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-utilities\") pod \"a4a9a67e-74d4-42bf-a563-555497bcd55c\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.322350 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-catalog-content\") pod \"a4a9a67e-74d4-42bf-a563-555497bcd55c\" (UID: \"a4a9a67e-74d4-42bf-a563-555497bcd55c\") " Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.323476 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-utilities" (OuterVolumeSpecName: "utilities") pod "a4a9a67e-74d4-42bf-a563-555497bcd55c" (UID: "a4a9a67e-74d4-42bf-a563-555497bcd55c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.329142 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4a9a67e-74d4-42bf-a563-555497bcd55c-kube-api-access-p9zt6" (OuterVolumeSpecName: "kube-api-access-p9zt6") pod "a4a9a67e-74d4-42bf-a563-555497bcd55c" (UID: "a4a9a67e-74d4-42bf-a563-555497bcd55c"). InnerVolumeSpecName "kube-api-access-p9zt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.424254 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.424291 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9zt6\" (UniqueName: \"kubernetes.io/projected/a4a9a67e-74d4-42bf-a563-555497bcd55c-kube-api-access-p9zt6\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.439552 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4a9a67e-74d4-42bf-a563-555497bcd55c" (UID: "a4a9a67e-74d4-42bf-a563-555497bcd55c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:51:33 crc kubenswrapper[4969]: I1004 08:51:33.526482 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9a67e-74d4-42bf-a563-555497bcd55c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:51:34 crc kubenswrapper[4969]: I1004 08:51:34.013116 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45pwx" event={"ID":"a4a9a67e-74d4-42bf-a563-555497bcd55c","Type":"ContainerDied","Data":"0f21b36a50c4fad5facce2283608a1b9d8651dd2e3b4740985bed843f20c949a"} Oct 04 08:51:34 crc kubenswrapper[4969]: I1004 08:51:34.013583 4969 scope.go:117] "RemoveContainer" containerID="c11814262f9039b86733dc8a0c4195c9efb64f7c044f622fa973be070ce5e568" Oct 04 08:51:34 crc kubenswrapper[4969]: I1004 08:51:34.013223 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-45pwx" Oct 04 08:51:34 crc kubenswrapper[4969]: I1004 08:51:34.052822 4969 scope.go:117] "RemoveContainer" containerID="8adb734b3cec2d57e79bc0958e996165b14ea64a109cfe0992db23a528451ece" Oct 04 08:51:34 crc kubenswrapper[4969]: I1004 08:51:34.072791 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-45pwx"] Oct 04 08:51:34 crc kubenswrapper[4969]: I1004 08:51:34.088031 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-45pwx"] Oct 04 08:51:34 crc kubenswrapper[4969]: I1004 08:51:34.089670 4969 scope.go:117] "RemoveContainer" containerID="1e5f746da655d545ec86577e59af0e58a481fd9062f6b9057ca7f65efa29aa4b" Oct 04 08:51:35 crc kubenswrapper[4969]: I1004 08:51:35.070412 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" path="/var/lib/kubelet/pods/a4a9a67e-74d4-42bf-a563-555497bcd55c/volumes" Oct 04 08:51:48 crc kubenswrapper[4969]: I1004 08:51:48.912861 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9qll4"] Oct 04 08:51:48 crc kubenswrapper[4969]: E1004 08:51:48.913909 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerName="registry-server" Oct 04 08:51:48 crc kubenswrapper[4969]: I1004 08:51:48.913926 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerName="registry-server" Oct 04 08:51:48 crc kubenswrapper[4969]: E1004 08:51:48.913940 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerName="extract-content" Oct 04 08:51:48 crc kubenswrapper[4969]: I1004 08:51:48.913946 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerName="extract-content" Oct 04 08:51:48 crc kubenswrapper[4969]: E1004 08:51:48.913968 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerName="extract-utilities" Oct 04 08:51:48 crc kubenswrapper[4969]: I1004 08:51:48.913975 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerName="extract-utilities" Oct 04 08:51:48 crc kubenswrapper[4969]: I1004 08:51:48.914207 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4a9a67e-74d4-42bf-a563-555497bcd55c" containerName="registry-server" Oct 04 08:51:48 crc kubenswrapper[4969]: I1004 08:51:48.916219 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:48 crc kubenswrapper[4969]: I1004 08:51:48.931523 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qll4"] Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.055671 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-utilities\") pod \"redhat-marketplace-9qll4\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.056018 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frpqf\" (UniqueName: \"kubernetes.io/projected/e46d31dd-813a-4d03-ae84-d9db8305e687-kube-api-access-frpqf\") pod \"redhat-marketplace-9qll4\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.056125 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-catalog-content\") pod \"redhat-marketplace-9qll4\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.158326 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-utilities\") pod \"redhat-marketplace-9qll4\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.158564 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frpqf\" (UniqueName: \"kubernetes.io/projected/e46d31dd-813a-4d03-ae84-d9db8305e687-kube-api-access-frpqf\") pod \"redhat-marketplace-9qll4\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.158624 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-catalog-content\") pod \"redhat-marketplace-9qll4\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.158976 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-utilities\") pod \"redhat-marketplace-9qll4\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.159138 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-catalog-content\") pod \"redhat-marketplace-9qll4\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.184561 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frpqf\" (UniqueName: \"kubernetes.io/projected/e46d31dd-813a-4d03-ae84-d9db8305e687-kube-api-access-frpqf\") pod \"redhat-marketplace-9qll4\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.250094 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:49 crc kubenswrapper[4969]: I1004 08:51:49.706933 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qll4"] Oct 04 08:51:50 crc kubenswrapper[4969]: I1004 08:51:50.187317 4969 generic.go:334] "Generic (PLEG): container finished" podID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerID="8bc3db1a8759e163eb44aecdce6f0583aea49f67253084e2971a8c94870dd347" exitCode=0 Oct 04 08:51:50 crc kubenswrapper[4969]: I1004 08:51:50.187704 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qll4" event={"ID":"e46d31dd-813a-4d03-ae84-d9db8305e687","Type":"ContainerDied","Data":"8bc3db1a8759e163eb44aecdce6f0583aea49f67253084e2971a8c94870dd347"} Oct 04 08:51:50 crc kubenswrapper[4969]: I1004 08:51:50.189874 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qll4" event={"ID":"e46d31dd-813a-4d03-ae84-d9db8305e687","Type":"ContainerStarted","Data":"3f1024818db259480f1dc63bfd0a76b01589d85fe4cbf62148f91f6a9f66320f"} Oct 04 08:51:52 crc kubenswrapper[4969]: I1004 08:51:52.214623 4969 generic.go:334] "Generic (PLEG): container finished" podID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerID="a472da62919fc6252170d3cbbcfd35f79612e27fb2b079f9db41350032eb8fcd" exitCode=0 Oct 04 08:51:52 crc kubenswrapper[4969]: I1004 08:51:52.214944 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qll4" event={"ID":"e46d31dd-813a-4d03-ae84-d9db8305e687","Type":"ContainerDied","Data":"a472da62919fc6252170d3cbbcfd35f79612e27fb2b079f9db41350032eb8fcd"} Oct 04 08:51:53 crc kubenswrapper[4969]: I1004 08:51:53.224667 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qll4" event={"ID":"e46d31dd-813a-4d03-ae84-d9db8305e687","Type":"ContainerStarted","Data":"9ddcfd14b21d406bbfacdfbbc8b25ae7e5313cff929aada96a53c5381cf6ec10"} Oct 04 08:51:53 crc kubenswrapper[4969]: I1004 08:51:53.247193 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9qll4" podStartSLOduration=2.681187216 podStartE2EDuration="5.247173679s" podCreationTimestamp="2025-10-04 08:51:48 +0000 UTC" firstStartedPulling="2025-10-04 08:51:50.192178233 +0000 UTC m=+2137.946447097" lastFinishedPulling="2025-10-04 08:51:52.758164746 +0000 UTC m=+2140.512433560" observedRunningTime="2025-10-04 08:51:53.243618272 +0000 UTC m=+2140.997887086" watchObservedRunningTime="2025-10-04 08:51:53.247173679 +0000 UTC m=+2141.001442493" Oct 04 08:51:59 crc kubenswrapper[4969]: I1004 08:51:59.250532 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:59 crc kubenswrapper[4969]: I1004 08:51:59.250996 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:59 crc kubenswrapper[4969]: I1004 08:51:59.304216 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:59 crc kubenswrapper[4969]: I1004 08:51:59.358261 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:51:59 crc kubenswrapper[4969]: I1004 08:51:59.549715 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qll4"] Oct 04 08:52:01 crc kubenswrapper[4969]: I1004 08:52:01.309825 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9qll4" podUID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerName="registry-server" containerID="cri-o://9ddcfd14b21d406bbfacdfbbc8b25ae7e5313cff929aada96a53c5381cf6ec10" gracePeriod=2 Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.321998 4969 generic.go:334] "Generic (PLEG): container finished" podID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerID="9ddcfd14b21d406bbfacdfbbc8b25ae7e5313cff929aada96a53c5381cf6ec10" exitCode=0 Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.322074 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qll4" event={"ID":"e46d31dd-813a-4d03-ae84-d9db8305e687","Type":"ContainerDied","Data":"9ddcfd14b21d406bbfacdfbbc8b25ae7e5313cff929aada96a53c5381cf6ec10"} Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.323558 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qll4" event={"ID":"e46d31dd-813a-4d03-ae84-d9db8305e687","Type":"ContainerDied","Data":"3f1024818db259480f1dc63bfd0a76b01589d85fe4cbf62148f91f6a9f66320f"} Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.323628 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f1024818db259480f1dc63bfd0a76b01589d85fe4cbf62148f91f6a9f66320f" Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.326625 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.447055 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frpqf\" (UniqueName: \"kubernetes.io/projected/e46d31dd-813a-4d03-ae84-d9db8305e687-kube-api-access-frpqf\") pod \"e46d31dd-813a-4d03-ae84-d9db8305e687\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.447139 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-utilities\") pod \"e46d31dd-813a-4d03-ae84-d9db8305e687\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.447262 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-catalog-content\") pod \"e46d31dd-813a-4d03-ae84-d9db8305e687\" (UID: \"e46d31dd-813a-4d03-ae84-d9db8305e687\") " Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.448325 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-utilities" (OuterVolumeSpecName: "utilities") pod "e46d31dd-813a-4d03-ae84-d9db8305e687" (UID: "e46d31dd-813a-4d03-ae84-d9db8305e687"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.453539 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e46d31dd-813a-4d03-ae84-d9db8305e687-kube-api-access-frpqf" (OuterVolumeSpecName: "kube-api-access-frpqf") pod "e46d31dd-813a-4d03-ae84-d9db8305e687" (UID: "e46d31dd-813a-4d03-ae84-d9db8305e687"). InnerVolumeSpecName "kube-api-access-frpqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.468034 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e46d31dd-813a-4d03-ae84-d9db8305e687" (UID: "e46d31dd-813a-4d03-ae84-d9db8305e687"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.549958 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frpqf\" (UniqueName: \"kubernetes.io/projected/e46d31dd-813a-4d03-ae84-d9db8305e687-kube-api-access-frpqf\") on node \"crc\" DevicePath \"\"" Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.549997 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:52:02 crc kubenswrapper[4969]: I1004 08:52:02.550019 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46d31dd-813a-4d03-ae84-d9db8305e687-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:52:03 crc kubenswrapper[4969]: I1004 08:52:03.336984 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qll4" Oct 04 08:52:03 crc kubenswrapper[4969]: I1004 08:52:03.366840 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qll4"] Oct 04 08:52:03 crc kubenswrapper[4969]: I1004 08:52:03.376539 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qll4"] Oct 04 08:52:05 crc kubenswrapper[4969]: I1004 08:52:05.071730 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e46d31dd-813a-4d03-ae84-d9db8305e687" path="/var/lib/kubelet/pods/e46d31dd-813a-4d03-ae84-d9db8305e687/volumes" Oct 04 08:52:33 crc kubenswrapper[4969]: I1004 08:52:33.683714 4969 generic.go:334] "Generic (PLEG): container finished" podID="ceaa1102-1d9c-40ec-a6a7-b31ca499c55e" containerID="a9fff868b811ac08167d559b3b3a21b7e22584d8860c0a4db72eff67465aae74" exitCode=0 Oct 04 08:52:33 crc kubenswrapper[4969]: I1004 08:52:33.683799 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" event={"ID":"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e","Type":"ContainerDied","Data":"a9fff868b811ac08167d559b3b3a21b7e22584d8860c0a4db72eff67465aae74"} Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.179713 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.350707 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovn-combined-ca-bundle\") pod \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.350892 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ssh-key\") pod \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.351046 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-inventory\") pod \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.351095 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovncontroller-config-0\") pod \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.351228 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfx4k\" (UniqueName: \"kubernetes.io/projected/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-kube-api-access-bfx4k\") pod \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\" (UID: \"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e\") " Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.357110 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-kube-api-access-bfx4k" (OuterVolumeSpecName: "kube-api-access-bfx4k") pod "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e" (UID: "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e"). InnerVolumeSpecName "kube-api-access-bfx4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.357866 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e" (UID: "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.379410 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-inventory" (OuterVolumeSpecName: "inventory") pod "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e" (UID: "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.381154 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e" (UID: "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.385900 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e" (UID: "ceaa1102-1d9c-40ec-a6a7-b31ca499c55e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.455205 4969 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.455259 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.455278 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.455298 4969 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.455489 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfx4k\" (UniqueName: \"kubernetes.io/projected/ceaa1102-1d9c-40ec-a6a7-b31ca499c55e-kube-api-access-bfx4k\") on node \"crc\" DevicePath \"\"" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.719455 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" event={"ID":"ceaa1102-1d9c-40ec-a6a7-b31ca499c55e","Type":"ContainerDied","Data":"3a6a2defa3245cfda56f2d45274087690d4753527d32ae05fc55931b2c1b3d29"} Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.719783 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a6a2defa3245cfda56f2d45274087690d4753527d32ae05fc55931b2c1b3d29" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.719720 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4hrwl" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.825792 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4"] Oct 04 08:52:35 crc kubenswrapper[4969]: E1004 08:52:35.826178 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceaa1102-1d9c-40ec-a6a7-b31ca499c55e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.826200 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceaa1102-1d9c-40ec-a6a7-b31ca499c55e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 08:52:35 crc kubenswrapper[4969]: E1004 08:52:35.826214 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerName="registry-server" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.826223 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerName="registry-server" Oct 04 08:52:35 crc kubenswrapper[4969]: E1004 08:52:35.826264 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerName="extract-utilities" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.826271 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerName="extract-utilities" Oct 04 08:52:35 crc kubenswrapper[4969]: E1004 08:52:35.826281 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerName="extract-content" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.826286 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerName="extract-content" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.826484 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceaa1102-1d9c-40ec-a6a7-b31ca499c55e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.826505 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="e46d31dd-813a-4d03-ae84-d9db8305e687" containerName="registry-server" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.827882 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.830258 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.830687 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.830918 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.831092 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.831166 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.831860 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.849340 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4"] Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.971376 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.971577 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.971632 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.971857 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js5w2\" (UniqueName: \"kubernetes.io/projected/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-kube-api-access-js5w2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.972297 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:35 crc kubenswrapper[4969]: I1004 08:52:35.972510 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.074901 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.075004 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.075044 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.075106 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.075141 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.075179 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js5w2\" (UniqueName: \"kubernetes.io/projected/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-kube-api-access-js5w2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.079764 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.081011 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.081504 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.082735 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.086105 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.100704 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js5w2\" (UniqueName: \"kubernetes.io/projected/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-kube-api-access-js5w2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.150177 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.711593 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4"] Oct 04 08:52:36 crc kubenswrapper[4969]: I1004 08:52:36.734309 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" event={"ID":"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6","Type":"ContainerStarted","Data":"c383175713f7283e83354d76d2c55be1a83b221f72d2a554f16a8e76efa7cabe"} Oct 04 08:52:37 crc kubenswrapper[4969]: I1004 08:52:37.747860 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" event={"ID":"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6","Type":"ContainerStarted","Data":"4639b2776302f55405d71ac12e6e7bdfd8ef22d37b4fc9a88d958fd6c3622fbd"} Oct 04 08:52:37 crc kubenswrapper[4969]: I1004 08:52:37.773492 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" podStartSLOduration=2.20986184 podStartE2EDuration="2.773473555s" podCreationTimestamp="2025-10-04 08:52:35 +0000 UTC" firstStartedPulling="2025-10-04 08:52:36.726827774 +0000 UTC m=+2184.481096628" lastFinishedPulling="2025-10-04 08:52:37.290439529 +0000 UTC m=+2185.044708343" observedRunningTime="2025-10-04 08:52:37.768663906 +0000 UTC m=+2185.522932730" watchObservedRunningTime="2025-10-04 08:52:37.773473555 +0000 UTC m=+2185.527742369" Oct 04 08:53:32 crc kubenswrapper[4969]: I1004 08:53:32.333150 4969 generic.go:334] "Generic (PLEG): container finished" podID="87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" containerID="4639b2776302f55405d71ac12e6e7bdfd8ef22d37b4fc9a88d958fd6c3622fbd" exitCode=0 Oct 04 08:53:32 crc kubenswrapper[4969]: I1004 08:53:32.333303 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" event={"ID":"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6","Type":"ContainerDied","Data":"4639b2776302f55405d71ac12e6e7bdfd8ef22d37b4fc9a88d958fd6c3622fbd"} Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.794533 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.951454 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js5w2\" (UniqueName: \"kubernetes.io/projected/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-kube-api-access-js5w2\") pod \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.951881 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-ssh-key\") pod \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.952003 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-inventory\") pod \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.952029 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-metadata-combined-ca-bundle\") pod \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.952090 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.952134 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-nova-metadata-neutron-config-0\") pod \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\" (UID: \"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6\") " Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.958168 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" (UID: "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.961844 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-kube-api-access-js5w2" (OuterVolumeSpecName: "kube-api-access-js5w2") pod "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" (UID: "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6"). InnerVolumeSpecName "kube-api-access-js5w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.993535 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" (UID: "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:53:33 crc kubenswrapper[4969]: I1004 08:53:33.995912 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" (UID: "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.005825 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" (UID: "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.005975 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-inventory" (OuterVolumeSpecName: "inventory") pod "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" (UID: "87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.054322 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.054397 4969 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.054463 4969 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.054487 4969 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.054503 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js5w2\" (UniqueName: \"kubernetes.io/projected/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-kube-api-access-js5w2\") on node \"crc\" DevicePath \"\"" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.054519 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.372223 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" event={"ID":"87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6","Type":"ContainerDied","Data":"c383175713f7283e83354d76d2c55be1a83b221f72d2a554f16a8e76efa7cabe"} Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.372269 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c383175713f7283e83354d76d2c55be1a83b221f72d2a554f16a8e76efa7cabe" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.372332 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.507984 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9"] Oct 04 08:53:34 crc kubenswrapper[4969]: E1004 08:53:34.508355 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.508374 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.508595 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.509243 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.511590 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.511774 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.512006 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.514218 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.514857 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.517856 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9"] Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.666009 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.666283 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.666311 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjljf\" (UniqueName: \"kubernetes.io/projected/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-kube-api-access-hjljf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.666345 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.666364 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.768455 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.768564 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.768612 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjljf\" (UniqueName: \"kubernetes.io/projected/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-kube-api-access-hjljf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.768704 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.768738 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.776400 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.776549 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.776710 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.777893 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.789466 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjljf\" (UniqueName: \"kubernetes.io/projected/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-kube-api-access-hjljf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:34 crc kubenswrapper[4969]: I1004 08:53:34.842204 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:53:35 crc kubenswrapper[4969]: I1004 08:53:35.419456 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9"] Oct 04 08:53:36 crc kubenswrapper[4969]: I1004 08:53:36.414609 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" event={"ID":"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb","Type":"ContainerStarted","Data":"9895f8bb6230a19efe796abdebfadfbb2eb9c5aa82054c44a17c819a07cb05f0"} Oct 04 08:53:36 crc kubenswrapper[4969]: I1004 08:53:36.415385 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" event={"ID":"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb","Type":"ContainerStarted","Data":"c885864086bde06aa73cc81d1f6f6e174f1d362cac3eda84fb709d166bb2476e"} Oct 04 08:53:36 crc kubenswrapper[4969]: I1004 08:53:36.442232 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" podStartSLOduration=1.84989831 podStartE2EDuration="2.442205962s" podCreationTimestamp="2025-10-04 08:53:34 +0000 UTC" firstStartedPulling="2025-10-04 08:53:35.421010289 +0000 UTC m=+2243.175279103" lastFinishedPulling="2025-10-04 08:53:36.013317941 +0000 UTC m=+2243.767586755" observedRunningTime="2025-10-04 08:53:36.438004789 +0000 UTC m=+2244.192273633" watchObservedRunningTime="2025-10-04 08:53:36.442205962 +0000 UTC m=+2244.196474796" Oct 04 08:53:49 crc kubenswrapper[4969]: I1004 08:53:49.666729 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:53:49 crc kubenswrapper[4969]: I1004 08:53:49.667746 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:54:19 crc kubenswrapper[4969]: I1004 08:54:19.666741 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:54:19 crc kubenswrapper[4969]: I1004 08:54:19.667435 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:54:49 crc kubenswrapper[4969]: I1004 08:54:49.666905 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 08:54:49 crc kubenswrapper[4969]: I1004 08:54:49.667622 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 08:54:49 crc kubenswrapper[4969]: I1004 08:54:49.667703 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 08:54:49 crc kubenswrapper[4969]: I1004 08:54:49.668878 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 08:54:49 crc kubenswrapper[4969]: I1004 08:54:49.669003 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" gracePeriod=600 Oct 04 08:54:49 crc kubenswrapper[4969]: E1004 08:54:49.799642 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:54:50 crc kubenswrapper[4969]: I1004 08:54:50.261278 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" exitCode=0 Oct 04 08:54:50 crc kubenswrapper[4969]: I1004 08:54:50.261318 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916"} Oct 04 08:54:50 crc kubenswrapper[4969]: I1004 08:54:50.261382 4969 scope.go:117] "RemoveContainer" containerID="f87372b3dbcc42c8848b2ff115cee7651dab7a8ca5bb7cad720f3bfdc85a0f10" Oct 04 08:54:50 crc kubenswrapper[4969]: I1004 08:54:50.262455 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:54:50 crc kubenswrapper[4969]: E1004 08:54:50.263085 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:55:02 crc kubenswrapper[4969]: I1004 08:55:02.054896 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:55:02 crc kubenswrapper[4969]: E1004 08:55:02.055874 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:55:13 crc kubenswrapper[4969]: I1004 08:55:13.066718 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:55:13 crc kubenswrapper[4969]: E1004 08:55:13.069097 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:55:25 crc kubenswrapper[4969]: I1004 08:55:25.055339 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:55:25 crc kubenswrapper[4969]: E1004 08:55:25.056122 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:55:32 crc kubenswrapper[4969]: I1004 08:55:32.394151 4969 scope.go:117] "RemoveContainer" containerID="07c68e0744c7c315a32c00f3639eabc232d85292a2af9fc167b6933d751fc3e4" Oct 04 08:55:32 crc kubenswrapper[4969]: I1004 08:55:32.429272 4969 scope.go:117] "RemoveContainer" containerID="1f554cd17f7d36ae623ed82b7a05641912a522c40e3fd8c27819dbba3d7d3b17" Oct 04 08:55:32 crc kubenswrapper[4969]: I1004 08:55:32.484006 4969 scope.go:117] "RemoveContainer" containerID="58ee703e670bd512e5339d1a7dbe758607034fa5ecd610df69e4609aced1e2a3" Oct 04 08:55:40 crc kubenswrapper[4969]: I1004 08:55:40.055455 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:55:40 crc kubenswrapper[4969]: E1004 08:55:40.056772 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:55:52 crc kubenswrapper[4969]: I1004 08:55:52.055838 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:55:52 crc kubenswrapper[4969]: E1004 08:55:52.056989 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:56:06 crc kubenswrapper[4969]: I1004 08:56:06.056041 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:56:06 crc kubenswrapper[4969]: E1004 08:56:06.056971 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:56:17 crc kubenswrapper[4969]: I1004 08:56:17.056550 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:56:17 crc kubenswrapper[4969]: E1004 08:56:17.058533 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:56:30 crc kubenswrapper[4969]: I1004 08:56:30.055995 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:56:30 crc kubenswrapper[4969]: E1004 08:56:30.056990 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:56:43 crc kubenswrapper[4969]: I1004 08:56:43.064263 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:56:43 crc kubenswrapper[4969]: E1004 08:56:43.072162 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:56:55 crc kubenswrapper[4969]: I1004 08:56:55.055763 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:56:55 crc kubenswrapper[4969]: E1004 08:56:55.056548 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:57:06 crc kubenswrapper[4969]: I1004 08:57:06.056535 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:57:06 crc kubenswrapper[4969]: E1004 08:57:06.057936 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:57:19 crc kubenswrapper[4969]: I1004 08:57:19.055974 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:57:19 crc kubenswrapper[4969]: E1004 08:57:19.057242 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:57:30 crc kubenswrapper[4969]: I1004 08:57:30.056415 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:57:30 crc kubenswrapper[4969]: E1004 08:57:30.057454 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:57:44 crc kubenswrapper[4969]: I1004 08:57:44.055019 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:57:44 crc kubenswrapper[4969]: E1004 08:57:44.055919 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:57:56 crc kubenswrapper[4969]: I1004 08:57:56.055972 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:57:56 crc kubenswrapper[4969]: E1004 08:57:56.057108 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:58:09 crc kubenswrapper[4969]: I1004 08:58:09.055503 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:58:09 crc kubenswrapper[4969]: E1004 08:58:09.056571 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:58:20 crc kubenswrapper[4969]: I1004 08:58:20.056013 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:58:20 crc kubenswrapper[4969]: E1004 08:58:20.057118 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:58:31 crc kubenswrapper[4969]: I1004 08:58:31.056445 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:58:31 crc kubenswrapper[4969]: E1004 08:58:31.057395 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:58:32 crc kubenswrapper[4969]: I1004 08:58:32.592903 4969 scope.go:117] "RemoveContainer" containerID="a472da62919fc6252170d3cbbcfd35f79612e27fb2b079f9db41350032eb8fcd" Oct 04 08:58:32 crc kubenswrapper[4969]: I1004 08:58:32.623532 4969 scope.go:117] "RemoveContainer" containerID="9ddcfd14b21d406bbfacdfbbc8b25ae7e5313cff929aada96a53c5381cf6ec10" Oct 04 08:58:32 crc kubenswrapper[4969]: I1004 08:58:32.712650 4969 scope.go:117] "RemoveContainer" containerID="8bc3db1a8759e163eb44aecdce6f0583aea49f67253084e2971a8c94870dd347" Oct 04 08:58:44 crc kubenswrapper[4969]: I1004 08:58:44.030511 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" event={"ID":"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb","Type":"ContainerDied","Data":"9895f8bb6230a19efe796abdebfadfbb2eb9c5aa82054c44a17c819a07cb05f0"} Oct 04 08:58:44 crc kubenswrapper[4969]: I1004 08:58:44.030526 4969 generic.go:334] "Generic (PLEG): container finished" podID="07ace2b0-20d8-4b9a-8f93-4649a3eee5bb" containerID="9895f8bb6230a19efe796abdebfadfbb2eb9c5aa82054c44a17c819a07cb05f0" exitCode=0 Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.056094 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:58:45 crc kubenswrapper[4969]: E1004 08:58:45.056713 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.499113 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.656287 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjljf\" (UniqueName: \"kubernetes.io/projected/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-kube-api-access-hjljf\") pod \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.656357 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-combined-ca-bundle\") pod \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.656482 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-secret-0\") pod \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.656671 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-ssh-key\") pod \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.656806 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-inventory\") pod \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\" (UID: \"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb\") " Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.662903 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-kube-api-access-hjljf" (OuterVolumeSpecName: "kube-api-access-hjljf") pod "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb" (UID: "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb"). InnerVolumeSpecName "kube-api-access-hjljf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.662919 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb" (UID: "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.692727 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-inventory" (OuterVolumeSpecName: "inventory") pod "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb" (UID: "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.694062 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb" (UID: "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.700654 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb" (UID: "07ace2b0-20d8-4b9a-8f93-4649a3eee5bb"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.759414 4969 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.759503 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.759517 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.759537 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjljf\" (UniqueName: \"kubernetes.io/projected/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-kube-api-access-hjljf\") on node \"crc\" DevicePath \"\"" Oct 04 08:58:45 crc kubenswrapper[4969]: I1004 08:58:45.759558 4969 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ace2b0-20d8-4b9a-8f93-4649a3eee5bb-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.053468 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" event={"ID":"07ace2b0-20d8-4b9a-8f93-4649a3eee5bb","Type":"ContainerDied","Data":"c885864086bde06aa73cc81d1f6f6e174f1d362cac3eda84fb709d166bb2476e"} Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.053528 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c885864086bde06aa73cc81d1f6f6e174f1d362cac3eda84fb709d166bb2476e" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.054011 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.195621 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f"] Oct 04 08:58:46 crc kubenswrapper[4969]: E1004 08:58:46.196831 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ace2b0-20d8-4b9a-8f93-4649a3eee5bb" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.196927 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ace2b0-20d8-4b9a-8f93-4649a3eee5bb" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.197239 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="07ace2b0-20d8-4b9a-8f93-4649a3eee5bb" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.198255 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.200679 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.200745 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.200753 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.200950 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.202034 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.202271 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.202803 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.221827 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f"] Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.373948 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.374117 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5p5h\" (UniqueName: \"kubernetes.io/projected/4ce3e260-2be8-435b-8a24-7d06614c0844-kube-api-access-m5p5h\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.374195 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.374350 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.374553 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.374621 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.374789 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.374935 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.374988 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.476109 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.476233 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.476281 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.476405 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.476550 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.476596 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.476704 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.476772 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5p5h\" (UniqueName: \"kubernetes.io/projected/4ce3e260-2be8-435b-8a24-7d06614c0844-kube-api-access-m5p5h\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.476820 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.478164 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.481899 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.485933 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.486126 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.486934 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.487839 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.488641 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.490298 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.497776 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5p5h\" (UniqueName: \"kubernetes.io/projected/4ce3e260-2be8-435b-8a24-7d06614c0844-kube-api-access-m5p5h\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p9v4f\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:46 crc kubenswrapper[4969]: I1004 08:58:46.531015 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 08:58:47 crc kubenswrapper[4969]: I1004 08:58:47.154196 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f"] Oct 04 08:58:47 crc kubenswrapper[4969]: W1004 08:58:47.162089 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ce3e260_2be8_435b_8a24_7d06614c0844.slice/crio-145235d3a6a698a650cfafdcb75f2d0ee1cead629ab389cca749dede1a945c40 WatchSource:0}: Error finding container 145235d3a6a698a650cfafdcb75f2d0ee1cead629ab389cca749dede1a945c40: Status 404 returned error can't find the container with id 145235d3a6a698a650cfafdcb75f2d0ee1cead629ab389cca749dede1a945c40 Oct 04 08:58:47 crc kubenswrapper[4969]: I1004 08:58:47.164828 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 08:58:48 crc kubenswrapper[4969]: I1004 08:58:48.077795 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" event={"ID":"4ce3e260-2be8-435b-8a24-7d06614c0844","Type":"ContainerStarted","Data":"2b40de07d62d74c6db7a19622db61a642550492fd4c51d91fea5a53fbc768b67"} Oct 04 08:58:48 crc kubenswrapper[4969]: I1004 08:58:48.078559 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" event={"ID":"4ce3e260-2be8-435b-8a24-7d06614c0844","Type":"ContainerStarted","Data":"145235d3a6a698a650cfafdcb75f2d0ee1cead629ab389cca749dede1a945c40"} Oct 04 08:58:48 crc kubenswrapper[4969]: I1004 08:58:48.102708 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" podStartSLOduration=1.5725367220000002 podStartE2EDuration="2.102689644s" podCreationTimestamp="2025-10-04 08:58:46 +0000 UTC" firstStartedPulling="2025-10-04 08:58:47.164395641 +0000 UTC m=+2554.918664485" lastFinishedPulling="2025-10-04 08:58:47.694548563 +0000 UTC m=+2555.448817407" observedRunningTime="2025-10-04 08:58:48.094465288 +0000 UTC m=+2555.848734132" watchObservedRunningTime="2025-10-04 08:58:48.102689644 +0000 UTC m=+2555.856958458" Oct 04 08:58:59 crc kubenswrapper[4969]: I1004 08:58:59.056003 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:58:59 crc kubenswrapper[4969]: E1004 08:58:59.057132 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:59:14 crc kubenswrapper[4969]: I1004 08:59:14.055565 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:59:14 crc kubenswrapper[4969]: E1004 08:59:14.057626 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:59:29 crc kubenswrapper[4969]: I1004 08:59:29.056097 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:59:29 crc kubenswrapper[4969]: E1004 08:59:29.056976 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.356169 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h47n5"] Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.361263 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.373298 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h47n5"] Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.473853 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddkn5\" (UniqueName: \"kubernetes.io/projected/57f7e121-aac1-4522-963b-282f096ed26c-kube-api-access-ddkn5\") pod \"certified-operators-h47n5\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.474075 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-utilities\") pod \"certified-operators-h47n5\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.474459 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-catalog-content\") pod \"certified-operators-h47n5\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.577026 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddkn5\" (UniqueName: \"kubernetes.io/projected/57f7e121-aac1-4522-963b-282f096ed26c-kube-api-access-ddkn5\") pod \"certified-operators-h47n5\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.577104 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-utilities\") pod \"certified-operators-h47n5\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.577269 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-catalog-content\") pod \"certified-operators-h47n5\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.577864 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-catalog-content\") pod \"certified-operators-h47n5\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.577929 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-utilities\") pod \"certified-operators-h47n5\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.598272 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddkn5\" (UniqueName: \"kubernetes.io/projected/57f7e121-aac1-4522-963b-282f096ed26c-kube-api-access-ddkn5\") pod \"certified-operators-h47n5\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:35 crc kubenswrapper[4969]: I1004 08:59:35.687719 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:36 crc kubenswrapper[4969]: I1004 08:59:36.233646 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h47n5"] Oct 04 08:59:36 crc kubenswrapper[4969]: I1004 08:59:36.629492 4969 generic.go:334] "Generic (PLEG): container finished" podID="57f7e121-aac1-4522-963b-282f096ed26c" containerID="a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145" exitCode=0 Oct 04 08:59:36 crc kubenswrapper[4969]: I1004 08:59:36.629545 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47n5" event={"ID":"57f7e121-aac1-4522-963b-282f096ed26c","Type":"ContainerDied","Data":"a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145"} Oct 04 08:59:36 crc kubenswrapper[4969]: I1004 08:59:36.629577 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47n5" event={"ID":"57f7e121-aac1-4522-963b-282f096ed26c","Type":"ContainerStarted","Data":"abef751e36e9216ac2dc9e922069168ad61faef52d916f8745d7739809861212"} Oct 04 08:59:38 crc kubenswrapper[4969]: I1004 08:59:38.661733 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47n5" event={"ID":"57f7e121-aac1-4522-963b-282f096ed26c","Type":"ContainerStarted","Data":"1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984"} Oct 04 08:59:39 crc kubenswrapper[4969]: I1004 08:59:39.674495 4969 generic.go:334] "Generic (PLEG): container finished" podID="57f7e121-aac1-4522-963b-282f096ed26c" containerID="1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984" exitCode=0 Oct 04 08:59:39 crc kubenswrapper[4969]: I1004 08:59:39.674557 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47n5" event={"ID":"57f7e121-aac1-4522-963b-282f096ed26c","Type":"ContainerDied","Data":"1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984"} Oct 04 08:59:40 crc kubenswrapper[4969]: I1004 08:59:40.688767 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47n5" event={"ID":"57f7e121-aac1-4522-963b-282f096ed26c","Type":"ContainerStarted","Data":"011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74"} Oct 04 08:59:41 crc kubenswrapper[4969]: I1004 08:59:41.056286 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:59:41 crc kubenswrapper[4969]: E1004 08:59:41.056711 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 08:59:41 crc kubenswrapper[4969]: I1004 08:59:41.743413 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h47n5" podStartSLOduration=3.052664411 podStartE2EDuration="6.743395605s" podCreationTimestamp="2025-10-04 08:59:35 +0000 UTC" firstStartedPulling="2025-10-04 08:59:36.63174432 +0000 UTC m=+2604.386013144" lastFinishedPulling="2025-10-04 08:59:40.322475524 +0000 UTC m=+2608.076744338" observedRunningTime="2025-10-04 08:59:41.739949418 +0000 UTC m=+2609.494218232" watchObservedRunningTime="2025-10-04 08:59:41.743395605 +0000 UTC m=+2609.497664419" Oct 04 08:59:45 crc kubenswrapper[4969]: I1004 08:59:45.689740 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:45 crc kubenswrapper[4969]: I1004 08:59:45.690381 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:45 crc kubenswrapper[4969]: I1004 08:59:45.758763 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:45 crc kubenswrapper[4969]: I1004 08:59:45.836207 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:46 crc kubenswrapper[4969]: I1004 08:59:46.000397 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h47n5"] Oct 04 08:59:47 crc kubenswrapper[4969]: I1004 08:59:47.781179 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h47n5" podUID="57f7e121-aac1-4522-963b-282f096ed26c" containerName="registry-server" containerID="cri-o://011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74" gracePeriod=2 Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.784389 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.797753 4969 generic.go:334] "Generic (PLEG): container finished" podID="57f7e121-aac1-4522-963b-282f096ed26c" containerID="011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74" exitCode=0 Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.797792 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47n5" event={"ID":"57f7e121-aac1-4522-963b-282f096ed26c","Type":"ContainerDied","Data":"011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74"} Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.797815 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h47n5" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.797861 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47n5" event={"ID":"57f7e121-aac1-4522-963b-282f096ed26c","Type":"ContainerDied","Data":"abef751e36e9216ac2dc9e922069168ad61faef52d916f8745d7739809861212"} Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.797884 4969 scope.go:117] "RemoveContainer" containerID="011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.848250 4969 scope.go:117] "RemoveContainer" containerID="1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.854113 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-catalog-content\") pod \"57f7e121-aac1-4522-963b-282f096ed26c\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.854292 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-utilities\") pod \"57f7e121-aac1-4522-963b-282f096ed26c\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.854329 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddkn5\" (UniqueName: \"kubernetes.io/projected/57f7e121-aac1-4522-963b-282f096ed26c-kube-api-access-ddkn5\") pod \"57f7e121-aac1-4522-963b-282f096ed26c\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.855109 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-utilities" (OuterVolumeSpecName: "utilities") pod "57f7e121-aac1-4522-963b-282f096ed26c" (UID: "57f7e121-aac1-4522-963b-282f096ed26c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.865309 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57f7e121-aac1-4522-963b-282f096ed26c-kube-api-access-ddkn5" (OuterVolumeSpecName: "kube-api-access-ddkn5") pod "57f7e121-aac1-4522-963b-282f096ed26c" (UID: "57f7e121-aac1-4522-963b-282f096ed26c"). InnerVolumeSpecName "kube-api-access-ddkn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.890573 4969 scope.go:117] "RemoveContainer" containerID="a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.955834 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.955865 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddkn5\" (UniqueName: \"kubernetes.io/projected/57f7e121-aac1-4522-963b-282f096ed26c-kube-api-access-ddkn5\") on node \"crc\" DevicePath \"\"" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.983043 4969 scope.go:117] "RemoveContainer" containerID="011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74" Oct 04 08:59:48 crc kubenswrapper[4969]: E1004 08:59:48.983469 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74\": container with ID starting with 011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74 not found: ID does not exist" containerID="011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.983501 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74"} err="failed to get container status \"011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74\": rpc error: code = NotFound desc = could not find container \"011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74\": container with ID starting with 011ae9623850f4792fd41b237d10653b18100bc4e6fd336410aa2d6aafd14f74 not found: ID does not exist" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.983525 4969 scope.go:117] "RemoveContainer" containerID="1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984" Oct 04 08:59:48 crc kubenswrapper[4969]: E1004 08:59:48.983924 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984\": container with ID starting with 1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984 not found: ID does not exist" containerID="1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.983973 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984"} err="failed to get container status \"1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984\": rpc error: code = NotFound desc = could not find container \"1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984\": container with ID starting with 1933151d76afcb4b2a88bec6e7dfbc81d3cb596963f62528b8fa2e8ac72e1984 not found: ID does not exist" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.984015 4969 scope.go:117] "RemoveContainer" containerID="a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145" Oct 04 08:59:48 crc kubenswrapper[4969]: E1004 08:59:48.984404 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145\": container with ID starting with a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145 not found: ID does not exist" containerID="a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145" Oct 04 08:59:48 crc kubenswrapper[4969]: I1004 08:59:48.984445 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145"} err="failed to get container status \"a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145\": rpc error: code = NotFound desc = could not find container \"a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145\": container with ID starting with a0cb6036e45dca1301b3a5230988297426a3d1af4ab5faab93ed2b54d15f8145 not found: ID does not exist" Oct 04 08:59:50 crc kubenswrapper[4969]: I1004 08:59:50.189337 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57f7e121-aac1-4522-963b-282f096ed26c" (UID: "57f7e121-aac1-4522-963b-282f096ed26c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:59:50 crc kubenswrapper[4969]: I1004 08:59:50.190613 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-catalog-content\") pod \"57f7e121-aac1-4522-963b-282f096ed26c\" (UID: \"57f7e121-aac1-4522-963b-282f096ed26c\") " Oct 04 08:59:50 crc kubenswrapper[4969]: W1004 08:59:50.190840 4969 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/57f7e121-aac1-4522-963b-282f096ed26c/volumes/kubernetes.io~empty-dir/catalog-content Oct 04 08:59:50 crc kubenswrapper[4969]: I1004 08:59:50.190855 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57f7e121-aac1-4522-963b-282f096ed26c" (UID: "57f7e121-aac1-4522-963b-282f096ed26c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 08:59:50 crc kubenswrapper[4969]: I1004 08:59:50.198394 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f7e121-aac1-4522-963b-282f096ed26c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 08:59:50 crc kubenswrapper[4969]: I1004 08:59:50.357974 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h47n5"] Oct 04 08:59:50 crc kubenswrapper[4969]: I1004 08:59:50.369621 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h47n5"] Oct 04 08:59:51 crc kubenswrapper[4969]: I1004 08:59:51.069379 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57f7e121-aac1-4522-963b-282f096ed26c" path="/var/lib/kubelet/pods/57f7e121-aac1-4522-963b-282f096ed26c/volumes" Oct 04 08:59:54 crc kubenswrapper[4969]: I1004 08:59:54.056986 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 08:59:54 crc kubenswrapper[4969]: I1004 08:59:54.864627 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"1de80285c1f273504a5d16302235acfdc134e2fc6fcbcc32f7101a1fbb2723c5"} Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.146400 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52"] Oct 04 09:00:00 crc kubenswrapper[4969]: E1004 09:00:00.147123 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f7e121-aac1-4522-963b-282f096ed26c" containerName="extract-utilities" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.147136 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f7e121-aac1-4522-963b-282f096ed26c" containerName="extract-utilities" Oct 04 09:00:00 crc kubenswrapper[4969]: E1004 09:00:00.147155 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f7e121-aac1-4522-963b-282f096ed26c" containerName="extract-content" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.147161 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f7e121-aac1-4522-963b-282f096ed26c" containerName="extract-content" Oct 04 09:00:00 crc kubenswrapper[4969]: E1004 09:00:00.147181 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f7e121-aac1-4522-963b-282f096ed26c" containerName="registry-server" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.147187 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f7e121-aac1-4522-963b-282f096ed26c" containerName="registry-server" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.147368 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="57f7e121-aac1-4522-963b-282f096ed26c" containerName="registry-server" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.148016 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.151699 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.158185 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.176708 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52"] Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.353339 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-config-volume\") pod \"collect-profiles-29326140-wbs52\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.353570 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7xxr\" (UniqueName: \"kubernetes.io/projected/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-kube-api-access-q7xxr\") pod \"collect-profiles-29326140-wbs52\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.353653 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-secret-volume\") pod \"collect-profiles-29326140-wbs52\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.455889 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7xxr\" (UniqueName: \"kubernetes.io/projected/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-kube-api-access-q7xxr\") pod \"collect-profiles-29326140-wbs52\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.456765 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-secret-volume\") pod \"collect-profiles-29326140-wbs52\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.456845 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-config-volume\") pod \"collect-profiles-29326140-wbs52\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.457740 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-config-volume\") pod \"collect-profiles-29326140-wbs52\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.467553 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-secret-volume\") pod \"collect-profiles-29326140-wbs52\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.477220 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7xxr\" (UniqueName: \"kubernetes.io/projected/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-kube-api-access-q7xxr\") pod \"collect-profiles-29326140-wbs52\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:00 crc kubenswrapper[4969]: I1004 09:00:00.770144 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:01 crc kubenswrapper[4969]: I1004 09:00:01.281253 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52"] Oct 04 09:00:01 crc kubenswrapper[4969]: W1004 09:00:01.293401 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d8d980d_5a47_45ac_bdc7_d881f07bf40d.slice/crio-aa89110d9f2ec374fca7cf9441565098a9a195009ea7a1aadbe8172330468362 WatchSource:0}: Error finding container aa89110d9f2ec374fca7cf9441565098a9a195009ea7a1aadbe8172330468362: Status 404 returned error can't find the container with id aa89110d9f2ec374fca7cf9441565098a9a195009ea7a1aadbe8172330468362 Oct 04 09:00:01 crc kubenswrapper[4969]: I1004 09:00:01.951502 4969 generic.go:334] "Generic (PLEG): container finished" podID="3d8d980d-5a47-45ac-bdc7-d881f07bf40d" containerID="12e3cd5901ff41c7811c65879f5f524bf40cd71dfd4579e18e310b7bf9f3b1ee" exitCode=0 Oct 04 09:00:01 crc kubenswrapper[4969]: I1004 09:00:01.951556 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" event={"ID":"3d8d980d-5a47-45ac-bdc7-d881f07bf40d","Type":"ContainerDied","Data":"12e3cd5901ff41c7811c65879f5f524bf40cd71dfd4579e18e310b7bf9f3b1ee"} Oct 04 09:00:01 crc kubenswrapper[4969]: I1004 09:00:01.951745 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" event={"ID":"3d8d980d-5a47-45ac-bdc7-d881f07bf40d","Type":"ContainerStarted","Data":"aa89110d9f2ec374fca7cf9441565098a9a195009ea7a1aadbe8172330468362"} Oct 04 09:00:03 crc kubenswrapper[4969]: I1004 09:00:03.949345 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:03 crc kubenswrapper[4969]: I1004 09:00:03.973831 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" event={"ID":"3d8d980d-5a47-45ac-bdc7-d881f07bf40d","Type":"ContainerDied","Data":"aa89110d9f2ec374fca7cf9441565098a9a195009ea7a1aadbe8172330468362"} Oct 04 09:00:03 crc kubenswrapper[4969]: I1004 09:00:03.973870 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa89110d9f2ec374fca7cf9441565098a9a195009ea7a1aadbe8172330468362" Oct 04 09:00:03 crc kubenswrapper[4969]: I1004 09:00:03.973924 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52" Oct 04 09:00:04 crc kubenswrapper[4969]: I1004 09:00:04.130635 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7xxr\" (UniqueName: \"kubernetes.io/projected/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-kube-api-access-q7xxr\") pod \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " Oct 04 09:00:04 crc kubenswrapper[4969]: I1004 09:00:04.130890 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-config-volume\") pod \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " Oct 04 09:00:04 crc kubenswrapper[4969]: I1004 09:00:04.130964 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-secret-volume\") pod \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\" (UID: \"3d8d980d-5a47-45ac-bdc7-d881f07bf40d\") " Oct 04 09:00:04 crc kubenswrapper[4969]: I1004 09:00:04.131900 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-config-volume" (OuterVolumeSpecName: "config-volume") pod "3d8d980d-5a47-45ac-bdc7-d881f07bf40d" (UID: "3d8d980d-5a47-45ac-bdc7-d881f07bf40d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 09:00:04 crc kubenswrapper[4969]: I1004 09:00:04.138206 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3d8d980d-5a47-45ac-bdc7-d881f07bf40d" (UID: "3d8d980d-5a47-45ac-bdc7-d881f07bf40d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:00:04 crc kubenswrapper[4969]: I1004 09:00:04.140202 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-kube-api-access-q7xxr" (OuterVolumeSpecName: "kube-api-access-q7xxr") pod "3d8d980d-5a47-45ac-bdc7-d881f07bf40d" (UID: "3d8d980d-5a47-45ac-bdc7-d881f07bf40d"). InnerVolumeSpecName "kube-api-access-q7xxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:00:04 crc kubenswrapper[4969]: I1004 09:00:04.235340 4969 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 09:00:04 crc kubenswrapper[4969]: I1004 09:00:04.235382 4969 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 09:00:04 crc kubenswrapper[4969]: I1004 09:00:04.235394 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7xxr\" (UniqueName: \"kubernetes.io/projected/3d8d980d-5a47-45ac-bdc7-d881f07bf40d-kube-api-access-q7xxr\") on node \"crc\" DevicePath \"\"" Oct 04 09:00:05 crc kubenswrapper[4969]: I1004 09:00:05.074070 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x"] Oct 04 09:00:05 crc kubenswrapper[4969]: I1004 09:00:05.089756 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326095-4p88x"] Oct 04 09:00:07 crc kubenswrapper[4969]: I1004 09:00:07.073053 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29195139-26ff-4158-ba2c-282b1b921ede" path="/var/lib/kubelet/pods/29195139-26ff-4158-ba2c-282b1b921ede/volumes" Oct 04 09:00:32 crc kubenswrapper[4969]: I1004 09:00:32.810384 4969 scope.go:117] "RemoveContainer" containerID="53985fcadbe8d55b3611ba05ae14b4440661d95f4a6cc1273dcc243bd82d6faa" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.155730 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29326141-knk55"] Oct 04 09:01:00 crc kubenswrapper[4969]: E1004 09:01:00.157057 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d8d980d-5a47-45ac-bdc7-d881f07bf40d" containerName="collect-profiles" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.157077 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d8d980d-5a47-45ac-bdc7-d881f07bf40d" containerName="collect-profiles" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.157321 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d8d980d-5a47-45ac-bdc7-d881f07bf40d" containerName="collect-profiles" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.158203 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.175184 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29326141-knk55"] Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.248235 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qfnt\" (UniqueName: \"kubernetes.io/projected/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-kube-api-access-5qfnt\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.248293 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-config-data\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.248338 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-fernet-keys\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.248380 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-combined-ca-bundle\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.350249 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qfnt\" (UniqueName: \"kubernetes.io/projected/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-kube-api-access-5qfnt\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.350307 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-config-data\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.350350 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-fernet-keys\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.350389 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-combined-ca-bundle\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.357907 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-fernet-keys\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.368158 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-combined-ca-bundle\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.368663 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-config-data\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.372661 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qfnt\" (UniqueName: \"kubernetes.io/projected/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-kube-api-access-5qfnt\") pod \"keystone-cron-29326141-knk55\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.489051 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:00 crc kubenswrapper[4969]: I1004 09:01:00.949972 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29326141-knk55"] Oct 04 09:01:01 crc kubenswrapper[4969]: I1004 09:01:01.727559 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326141-knk55" event={"ID":"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a","Type":"ContainerStarted","Data":"312b1bb98ad2d9b8611d9f351148153cb2204b5e616941c073460a6f2d6e9675"} Oct 04 09:01:01 crc kubenswrapper[4969]: I1004 09:01:01.728092 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326141-knk55" event={"ID":"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a","Type":"ContainerStarted","Data":"700ced4ea8da7f1ca87274ca5c211c073279824fb276f9d82168847a1d5c57cc"} Oct 04 09:01:01 crc kubenswrapper[4969]: I1004 09:01:01.752499 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29326141-knk55" podStartSLOduration=1.752481023 podStartE2EDuration="1.752481023s" podCreationTimestamp="2025-10-04 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 09:01:01.745465281 +0000 UTC m=+2689.499734135" watchObservedRunningTime="2025-10-04 09:01:01.752481023 +0000 UTC m=+2689.506749837" Oct 04 09:01:03 crc kubenswrapper[4969]: I1004 09:01:03.751599 4969 generic.go:334] "Generic (PLEG): container finished" podID="cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a" containerID="312b1bb98ad2d9b8611d9f351148153cb2204b5e616941c073460a6f2d6e9675" exitCode=0 Oct 04 09:01:03 crc kubenswrapper[4969]: I1004 09:01:03.751770 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326141-knk55" event={"ID":"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a","Type":"ContainerDied","Data":"312b1bb98ad2d9b8611d9f351148153cb2204b5e616941c073460a6f2d6e9675"} Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.127337 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.246765 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-config-data\") pod \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.247032 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-combined-ca-bundle\") pod \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.247101 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-fernet-keys\") pod \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.247158 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qfnt\" (UniqueName: \"kubernetes.io/projected/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-kube-api-access-5qfnt\") pod \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\" (UID: \"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a\") " Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.254501 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-kube-api-access-5qfnt" (OuterVolumeSpecName: "kube-api-access-5qfnt") pod "cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a" (UID: "cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a"). InnerVolumeSpecName "kube-api-access-5qfnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.255554 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a" (UID: "cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.284648 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a" (UID: "cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.332919 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-config-data" (OuterVolumeSpecName: "config-data") pod "cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a" (UID: "cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.349705 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.349752 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.349770 4969 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.349783 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qfnt\" (UniqueName: \"kubernetes.io/projected/cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a-kube-api-access-5qfnt\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.775253 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326141-knk55" event={"ID":"cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a","Type":"ContainerDied","Data":"700ced4ea8da7f1ca87274ca5c211c073279824fb276f9d82168847a1d5c57cc"} Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.775309 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="700ced4ea8da7f1ca87274ca5c211c073279824fb276f9d82168847a1d5c57cc" Oct 04 09:01:05 crc kubenswrapper[4969]: I1004 09:01:05.775328 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326141-knk55" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.803041 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g8fsz"] Oct 04 09:01:14 crc kubenswrapper[4969]: E1004 09:01:14.804372 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a" containerName="keystone-cron" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.804404 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a" containerName="keystone-cron" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.804925 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a" containerName="keystone-cron" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.807609 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.847249 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g8fsz"] Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.886756 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-catalog-content\") pod \"community-operators-g8fsz\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.887013 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hwwp\" (UniqueName: \"kubernetes.io/projected/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-kube-api-access-9hwwp\") pod \"community-operators-g8fsz\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.887051 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-utilities\") pod \"community-operators-g8fsz\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.988723 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hwwp\" (UniqueName: \"kubernetes.io/projected/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-kube-api-access-9hwwp\") pod \"community-operators-g8fsz\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.988785 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-utilities\") pod \"community-operators-g8fsz\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.988817 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-catalog-content\") pod \"community-operators-g8fsz\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.989408 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-catalog-content\") pod \"community-operators-g8fsz\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:14 crc kubenswrapper[4969]: I1004 09:01:14.989839 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-utilities\") pod \"community-operators-g8fsz\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:15 crc kubenswrapper[4969]: I1004 09:01:15.011598 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hwwp\" (UniqueName: \"kubernetes.io/projected/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-kube-api-access-9hwwp\") pod \"community-operators-g8fsz\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:15 crc kubenswrapper[4969]: I1004 09:01:15.137049 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:15 crc kubenswrapper[4969]: I1004 09:01:15.627989 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g8fsz"] Oct 04 09:01:15 crc kubenswrapper[4969]: I1004 09:01:15.884459 4969 generic.go:334] "Generic (PLEG): container finished" podID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerID="7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68" exitCode=0 Oct 04 09:01:15 crc kubenswrapper[4969]: I1004 09:01:15.884502 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8fsz" event={"ID":"7f24a8b5-3039-49f2-ad6a-d2a24c79b300","Type":"ContainerDied","Data":"7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68"} Oct 04 09:01:15 crc kubenswrapper[4969]: I1004 09:01:15.885557 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8fsz" event={"ID":"7f24a8b5-3039-49f2-ad6a-d2a24c79b300","Type":"ContainerStarted","Data":"bd6a24c176256146061a57d9038b1fa677a195b6f0d329df42789187380c7194"} Oct 04 09:01:16 crc kubenswrapper[4969]: I1004 09:01:16.904934 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8fsz" event={"ID":"7f24a8b5-3039-49f2-ad6a-d2a24c79b300","Type":"ContainerStarted","Data":"63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f"} Oct 04 09:01:17 crc kubenswrapper[4969]: I1004 09:01:17.917482 4969 generic.go:334] "Generic (PLEG): container finished" podID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerID="63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f" exitCode=0 Oct 04 09:01:17 crc kubenswrapper[4969]: I1004 09:01:17.917629 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8fsz" event={"ID":"7f24a8b5-3039-49f2-ad6a-d2a24c79b300","Type":"ContainerDied","Data":"63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f"} Oct 04 09:01:18 crc kubenswrapper[4969]: I1004 09:01:18.930825 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8fsz" event={"ID":"7f24a8b5-3039-49f2-ad6a-d2a24c79b300","Type":"ContainerStarted","Data":"f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7"} Oct 04 09:01:18 crc kubenswrapper[4969]: I1004 09:01:18.986581 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g8fsz" podStartSLOduration=2.5086922080000003 podStartE2EDuration="4.986554815s" podCreationTimestamp="2025-10-04 09:01:14 +0000 UTC" firstStartedPulling="2025-10-04 09:01:15.886106366 +0000 UTC m=+2703.640375180" lastFinishedPulling="2025-10-04 09:01:18.363968963 +0000 UTC m=+2706.118237787" observedRunningTime="2025-10-04 09:01:18.9704545 +0000 UTC m=+2706.724723314" watchObservedRunningTime="2025-10-04 09:01:18.986554815 +0000 UTC m=+2706.740823629" Oct 04 09:01:25 crc kubenswrapper[4969]: I1004 09:01:25.138696 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:25 crc kubenswrapper[4969]: I1004 09:01:25.139308 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:25 crc kubenswrapper[4969]: I1004 09:01:25.227911 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:26 crc kubenswrapper[4969]: I1004 09:01:26.103697 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:26 crc kubenswrapper[4969]: I1004 09:01:26.157624 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g8fsz"] Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.046769 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g8fsz" podUID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerName="registry-server" containerID="cri-o://f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7" gracePeriod=2 Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.555143 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.712760 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hwwp\" (UniqueName: \"kubernetes.io/projected/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-kube-api-access-9hwwp\") pod \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.712981 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-catalog-content\") pod \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.713133 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-utilities\") pod \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\" (UID: \"7f24a8b5-3039-49f2-ad6a-d2a24c79b300\") " Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.713872 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-utilities" (OuterVolumeSpecName: "utilities") pod "7f24a8b5-3039-49f2-ad6a-d2a24c79b300" (UID: "7f24a8b5-3039-49f2-ad6a-d2a24c79b300"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.717484 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-kube-api-access-9hwwp" (OuterVolumeSpecName: "kube-api-access-9hwwp") pod "7f24a8b5-3039-49f2-ad6a-d2a24c79b300" (UID: "7f24a8b5-3039-49f2-ad6a-d2a24c79b300"). InnerVolumeSpecName "kube-api-access-9hwwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.775674 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f24a8b5-3039-49f2-ad6a-d2a24c79b300" (UID: "7f24a8b5-3039-49f2-ad6a-d2a24c79b300"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.815987 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.816022 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hwwp\" (UniqueName: \"kubernetes.io/projected/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-kube-api-access-9hwwp\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:28 crc kubenswrapper[4969]: I1004 09:01:28.816036 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f24a8b5-3039-49f2-ad6a-d2a24c79b300-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.057829 4969 generic.go:334] "Generic (PLEG): container finished" podID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerID="f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7" exitCode=0 Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.057893 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8fsz" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.067706 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8fsz" event={"ID":"7f24a8b5-3039-49f2-ad6a-d2a24c79b300","Type":"ContainerDied","Data":"f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7"} Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.067778 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8fsz" event={"ID":"7f24a8b5-3039-49f2-ad6a-d2a24c79b300","Type":"ContainerDied","Data":"bd6a24c176256146061a57d9038b1fa677a195b6f0d329df42789187380c7194"} Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.067821 4969 scope.go:117] "RemoveContainer" containerID="f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.102777 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g8fsz"] Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.106767 4969 scope.go:117] "RemoveContainer" containerID="63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.116523 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g8fsz"] Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.136274 4969 scope.go:117] "RemoveContainer" containerID="7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.185924 4969 scope.go:117] "RemoveContainer" containerID="f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7" Oct 04 09:01:29 crc kubenswrapper[4969]: E1004 09:01:29.186401 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7\": container with ID starting with f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7 not found: ID does not exist" containerID="f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.186485 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7"} err="failed to get container status \"f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7\": rpc error: code = NotFound desc = could not find container \"f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7\": container with ID starting with f5862ef75e61bc838faee432aaa6dae0ab37b68f30e4e5f1876b3d5d06ea5be7 not found: ID does not exist" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.186557 4969 scope.go:117] "RemoveContainer" containerID="63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f" Oct 04 09:01:29 crc kubenswrapper[4969]: E1004 09:01:29.187070 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f\": container with ID starting with 63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f not found: ID does not exist" containerID="63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.187133 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f"} err="failed to get container status \"63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f\": rpc error: code = NotFound desc = could not find container \"63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f\": container with ID starting with 63e17b003fc83da69dd1addd8a82f7dda48bbadb14c0bb8825076a5257b9b91f not found: ID does not exist" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.187155 4969 scope.go:117] "RemoveContainer" containerID="7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68" Oct 04 09:01:29 crc kubenswrapper[4969]: E1004 09:01:29.187408 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68\": container with ID starting with 7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68 not found: ID does not exist" containerID="7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.187476 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68"} err="failed to get container status \"7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68\": rpc error: code = NotFound desc = could not find container \"7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68\": container with ID starting with 7c3e2273791cc4cf84484befcd36454c0c5c700788adc419c2e660fcd8e82d68 not found: ID does not exist" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.899282 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qn7lg"] Oct 04 09:01:29 crc kubenswrapper[4969]: E1004 09:01:29.899993 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerName="registry-server" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.900024 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerName="registry-server" Oct 04 09:01:29 crc kubenswrapper[4969]: E1004 09:01:29.900082 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerName="extract-utilities" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.900096 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerName="extract-utilities" Oct 04 09:01:29 crc kubenswrapper[4969]: E1004 09:01:29.900132 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerName="extract-content" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.900148 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerName="extract-content" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.900544 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" containerName="registry-server" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.906295 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:29 crc kubenswrapper[4969]: I1004 09:01:29.915629 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qn7lg"] Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.041764 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvtvc\" (UniqueName: \"kubernetes.io/projected/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-kube-api-access-fvtvc\") pod \"redhat-operators-qn7lg\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.042060 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-utilities\") pod \"redhat-operators-qn7lg\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.042716 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-catalog-content\") pod \"redhat-operators-qn7lg\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.144769 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-catalog-content\") pod \"redhat-operators-qn7lg\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.145012 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvtvc\" (UniqueName: \"kubernetes.io/projected/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-kube-api-access-fvtvc\") pod \"redhat-operators-qn7lg\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.145060 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-utilities\") pod \"redhat-operators-qn7lg\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.145517 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-catalog-content\") pod \"redhat-operators-qn7lg\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.145537 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-utilities\") pod \"redhat-operators-qn7lg\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.183868 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvtvc\" (UniqueName: \"kubernetes.io/projected/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-kube-api-access-fvtvc\") pod \"redhat-operators-qn7lg\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.238310 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:30 crc kubenswrapper[4969]: I1004 09:01:30.714061 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qn7lg"] Oct 04 09:01:31 crc kubenswrapper[4969]: I1004 09:01:31.065010 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f24a8b5-3039-49f2-ad6a-d2a24c79b300" path="/var/lib/kubelet/pods/7f24a8b5-3039-49f2-ad6a-d2a24c79b300/volumes" Oct 04 09:01:31 crc kubenswrapper[4969]: I1004 09:01:31.078619 4969 generic.go:334] "Generic (PLEG): container finished" podID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerID="a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8" exitCode=0 Oct 04 09:01:31 crc kubenswrapper[4969]: I1004 09:01:31.078663 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn7lg" event={"ID":"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6","Type":"ContainerDied","Data":"a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8"} Oct 04 09:01:31 crc kubenswrapper[4969]: I1004 09:01:31.078685 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn7lg" event={"ID":"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6","Type":"ContainerStarted","Data":"02917fdf9c555ec3448636b28834e00477cbea9c45e42bbdd74f95b38322d554"} Oct 04 09:01:33 crc kubenswrapper[4969]: I1004 09:01:33.113392 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn7lg" event={"ID":"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6","Type":"ContainerStarted","Data":"eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c"} Oct 04 09:01:35 crc kubenswrapper[4969]: I1004 09:01:35.137473 4969 generic.go:334] "Generic (PLEG): container finished" podID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerID="eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c" exitCode=0 Oct 04 09:01:35 crc kubenswrapper[4969]: I1004 09:01:35.137537 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn7lg" event={"ID":"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6","Type":"ContainerDied","Data":"eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c"} Oct 04 09:01:36 crc kubenswrapper[4969]: I1004 09:01:36.151252 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn7lg" event={"ID":"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6","Type":"ContainerStarted","Data":"9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999"} Oct 04 09:01:36 crc kubenswrapper[4969]: I1004 09:01:36.182683 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qn7lg" podStartSLOduration=2.6405484660000003 podStartE2EDuration="7.182660904s" podCreationTimestamp="2025-10-04 09:01:29 +0000 UTC" firstStartedPulling="2025-10-04 09:01:31.080190443 +0000 UTC m=+2718.834459257" lastFinishedPulling="2025-10-04 09:01:35.622302871 +0000 UTC m=+2723.376571695" observedRunningTime="2025-10-04 09:01:36.176937273 +0000 UTC m=+2723.931206107" watchObservedRunningTime="2025-10-04 09:01:36.182660904 +0000 UTC m=+2723.936929728" Oct 04 09:01:40 crc kubenswrapper[4969]: I1004 09:01:40.239515 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:40 crc kubenswrapper[4969]: I1004 09:01:40.239796 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:41 crc kubenswrapper[4969]: I1004 09:01:41.301281 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qn7lg" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerName="registry-server" probeResult="failure" output=< Oct 04 09:01:41 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:01:41 crc kubenswrapper[4969]: > Oct 04 09:01:50 crc kubenswrapper[4969]: I1004 09:01:50.330913 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:50 crc kubenswrapper[4969]: I1004 09:01:50.424956 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:50 crc kubenswrapper[4969]: I1004 09:01:50.577951 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qn7lg"] Oct 04 09:01:52 crc kubenswrapper[4969]: I1004 09:01:52.353592 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qn7lg" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerName="registry-server" containerID="cri-o://9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999" gracePeriod=2 Oct 04 09:01:52 crc kubenswrapper[4969]: I1004 09:01:52.884124 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:52 crc kubenswrapper[4969]: I1004 09:01:52.962387 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvtvc\" (UniqueName: \"kubernetes.io/projected/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-kube-api-access-fvtvc\") pod \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " Oct 04 09:01:52 crc kubenswrapper[4969]: I1004 09:01:52.962966 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-catalog-content\") pod \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " Oct 04 09:01:52 crc kubenswrapper[4969]: I1004 09:01:52.963135 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-utilities\") pod \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\" (UID: \"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6\") " Oct 04 09:01:52 crc kubenswrapper[4969]: I1004 09:01:52.964281 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-utilities" (OuterVolumeSpecName: "utilities") pod "6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" (UID: "6cc395e0-2ea2-4f24-a114-47a17c1a3aa6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:01:52 crc kubenswrapper[4969]: I1004 09:01:52.970116 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-kube-api-access-fvtvc" (OuterVolumeSpecName: "kube-api-access-fvtvc") pod "6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" (UID: "6cc395e0-2ea2-4f24-a114-47a17c1a3aa6"). InnerVolumeSpecName "kube-api-access-fvtvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.066935 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.067020 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvtvc\" (UniqueName: \"kubernetes.io/projected/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-kube-api-access-fvtvc\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.068550 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" (UID: "6cc395e0-2ea2-4f24-a114-47a17c1a3aa6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.169146 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.370794 4969 generic.go:334] "Generic (PLEG): container finished" podID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerID="9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999" exitCode=0 Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.370899 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn7lg" event={"ID":"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6","Type":"ContainerDied","Data":"9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999"} Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.370952 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn7lg" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.370994 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn7lg" event={"ID":"6cc395e0-2ea2-4f24-a114-47a17c1a3aa6","Type":"ContainerDied","Data":"02917fdf9c555ec3448636b28834e00477cbea9c45e42bbdd74f95b38322d554"} Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.371031 4969 scope.go:117] "RemoveContainer" containerID="9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.407478 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qn7lg"] Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.420131 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qn7lg"] Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.424783 4969 scope.go:117] "RemoveContainer" containerID="eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.454229 4969 scope.go:117] "RemoveContainer" containerID="a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.501582 4969 scope.go:117] "RemoveContainer" containerID="9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999" Oct 04 09:01:53 crc kubenswrapper[4969]: E1004 09:01:53.502306 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999\": container with ID starting with 9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999 not found: ID does not exist" containerID="9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.502360 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999"} err="failed to get container status \"9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999\": rpc error: code = NotFound desc = could not find container \"9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999\": container with ID starting with 9ed672ec753c2c0f8da55ed4066283a854f0fd4e8e230e22778e1a6253440999 not found: ID does not exist" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.502410 4969 scope.go:117] "RemoveContainer" containerID="eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c" Oct 04 09:01:53 crc kubenswrapper[4969]: E1004 09:01:53.502998 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c\": container with ID starting with eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c not found: ID does not exist" containerID="eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.503046 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c"} err="failed to get container status \"eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c\": rpc error: code = NotFound desc = could not find container \"eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c\": container with ID starting with eaf10abb882d21765fa2b392e040a011d5bbb6f1f866ed3dcdfd30ef041a143c not found: ID does not exist" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.503080 4969 scope.go:117] "RemoveContainer" containerID="a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8" Oct 04 09:01:53 crc kubenswrapper[4969]: E1004 09:01:53.503394 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8\": container with ID starting with a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8 not found: ID does not exist" containerID="a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8" Oct 04 09:01:53 crc kubenswrapper[4969]: I1004 09:01:53.503439 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8"} err="failed to get container status \"a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8\": rpc error: code = NotFound desc = could not find container \"a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8\": container with ID starting with a3936c12bf0af82958be794d235120e664a8a86f71fa1dc3d058c21446d150d8 not found: ID does not exist" Oct 04 09:01:55 crc kubenswrapper[4969]: I1004 09:01:55.072611 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" path="/var/lib/kubelet/pods/6cc395e0-2ea2-4f24-a114-47a17c1a3aa6/volumes" Oct 04 09:02:19 crc kubenswrapper[4969]: I1004 09:02:19.667064 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:02:19 crc kubenswrapper[4969]: I1004 09:02:19.667597 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.133933 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vp7lv"] Oct 04 09:02:25 crc kubenswrapper[4969]: E1004 09:02:25.135270 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerName="extract-utilities" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.135294 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerName="extract-utilities" Oct 04 09:02:25 crc kubenswrapper[4969]: E1004 09:02:25.135317 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerName="registry-server" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.135329 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerName="registry-server" Oct 04 09:02:25 crc kubenswrapper[4969]: E1004 09:02:25.135350 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerName="extract-content" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.135364 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerName="extract-content" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.135765 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cc395e0-2ea2-4f24-a114-47a17c1a3aa6" containerName="registry-server" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.138341 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.147688 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vp7lv"] Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.203679 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-catalog-content\") pod \"redhat-marketplace-vp7lv\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.203837 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-utilities\") pod \"redhat-marketplace-vp7lv\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.203899 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb6f4\" (UniqueName: \"kubernetes.io/projected/824e57b5-550c-4c24-b243-b08c20be4a40-kube-api-access-kb6f4\") pod \"redhat-marketplace-vp7lv\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.305165 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-catalog-content\") pod \"redhat-marketplace-vp7lv\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.305335 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-utilities\") pod \"redhat-marketplace-vp7lv\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.305409 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb6f4\" (UniqueName: \"kubernetes.io/projected/824e57b5-550c-4c24-b243-b08c20be4a40-kube-api-access-kb6f4\") pod \"redhat-marketplace-vp7lv\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.305642 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-catalog-content\") pod \"redhat-marketplace-vp7lv\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.305744 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-utilities\") pod \"redhat-marketplace-vp7lv\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.324679 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb6f4\" (UniqueName: \"kubernetes.io/projected/824e57b5-550c-4c24-b243-b08c20be4a40-kube-api-access-kb6f4\") pod \"redhat-marketplace-vp7lv\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.485690 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:25 crc kubenswrapper[4969]: I1004 09:02:25.766962 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vp7lv"] Oct 04 09:02:26 crc kubenswrapper[4969]: I1004 09:02:26.783328 4969 generic.go:334] "Generic (PLEG): container finished" podID="824e57b5-550c-4c24-b243-b08c20be4a40" containerID="4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96" exitCode=0 Oct 04 09:02:26 crc kubenswrapper[4969]: I1004 09:02:26.783455 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp7lv" event={"ID":"824e57b5-550c-4c24-b243-b08c20be4a40","Type":"ContainerDied","Data":"4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96"} Oct 04 09:02:26 crc kubenswrapper[4969]: I1004 09:02:26.784134 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp7lv" event={"ID":"824e57b5-550c-4c24-b243-b08c20be4a40","Type":"ContainerStarted","Data":"7f6b2f0f0b994456eda60aec1a77098ebcebab73bf8a734ec2a0ee35025e7d66"} Oct 04 09:02:28 crc kubenswrapper[4969]: I1004 09:02:28.811717 4969 generic.go:334] "Generic (PLEG): container finished" podID="824e57b5-550c-4c24-b243-b08c20be4a40" containerID="9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9" exitCode=0 Oct 04 09:02:28 crc kubenswrapper[4969]: I1004 09:02:28.811798 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp7lv" event={"ID":"824e57b5-550c-4c24-b243-b08c20be4a40","Type":"ContainerDied","Data":"9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9"} Oct 04 09:02:29 crc kubenswrapper[4969]: I1004 09:02:29.824904 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp7lv" event={"ID":"824e57b5-550c-4c24-b243-b08c20be4a40","Type":"ContainerStarted","Data":"7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f"} Oct 04 09:02:29 crc kubenswrapper[4969]: I1004 09:02:29.850995 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vp7lv" podStartSLOduration=2.2984518449999998 podStartE2EDuration="4.850972886s" podCreationTimestamp="2025-10-04 09:02:25 +0000 UTC" firstStartedPulling="2025-10-04 09:02:26.785673451 +0000 UTC m=+2774.539942275" lastFinishedPulling="2025-10-04 09:02:29.338194462 +0000 UTC m=+2777.092463316" observedRunningTime="2025-10-04 09:02:29.846785203 +0000 UTC m=+2777.601054027" watchObservedRunningTime="2025-10-04 09:02:29.850972886 +0000 UTC m=+2777.605241720" Oct 04 09:02:35 crc kubenswrapper[4969]: I1004 09:02:35.485848 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:35 crc kubenswrapper[4969]: I1004 09:02:35.486497 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:35 crc kubenswrapper[4969]: I1004 09:02:35.572749 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:35 crc kubenswrapper[4969]: I1004 09:02:35.931399 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:35 crc kubenswrapper[4969]: I1004 09:02:35.976963 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vp7lv"] Oct 04 09:02:37 crc kubenswrapper[4969]: I1004 09:02:37.896329 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vp7lv" podUID="824e57b5-550c-4c24-b243-b08c20be4a40" containerName="registry-server" containerID="cri-o://7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f" gracePeriod=2 Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.392661 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.574208 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kb6f4\" (UniqueName: \"kubernetes.io/projected/824e57b5-550c-4c24-b243-b08c20be4a40-kube-api-access-kb6f4\") pod \"824e57b5-550c-4c24-b243-b08c20be4a40\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.574485 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-utilities\") pod \"824e57b5-550c-4c24-b243-b08c20be4a40\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.574538 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-catalog-content\") pod \"824e57b5-550c-4c24-b243-b08c20be4a40\" (UID: \"824e57b5-550c-4c24-b243-b08c20be4a40\") " Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.575262 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-utilities" (OuterVolumeSpecName: "utilities") pod "824e57b5-550c-4c24-b243-b08c20be4a40" (UID: "824e57b5-550c-4c24-b243-b08c20be4a40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.585629 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/824e57b5-550c-4c24-b243-b08c20be4a40-kube-api-access-kb6f4" (OuterVolumeSpecName: "kube-api-access-kb6f4") pod "824e57b5-550c-4c24-b243-b08c20be4a40" (UID: "824e57b5-550c-4c24-b243-b08c20be4a40"). InnerVolumeSpecName "kube-api-access-kb6f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.586624 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "824e57b5-550c-4c24-b243-b08c20be4a40" (UID: "824e57b5-550c-4c24-b243-b08c20be4a40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.677475 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.677715 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824e57b5-550c-4c24-b243-b08c20be4a40-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.677782 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kb6f4\" (UniqueName: \"kubernetes.io/projected/824e57b5-550c-4c24-b243-b08c20be4a40-kube-api-access-kb6f4\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.905615 4969 generic.go:334] "Generic (PLEG): container finished" podID="824e57b5-550c-4c24-b243-b08c20be4a40" containerID="7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f" exitCode=0 Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.905654 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp7lv" event={"ID":"824e57b5-550c-4c24-b243-b08c20be4a40","Type":"ContainerDied","Data":"7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f"} Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.905678 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp7lv" event={"ID":"824e57b5-550c-4c24-b243-b08c20be4a40","Type":"ContainerDied","Data":"7f6b2f0f0b994456eda60aec1a77098ebcebab73bf8a734ec2a0ee35025e7d66"} Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.905696 4969 scope.go:117] "RemoveContainer" containerID="7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.906734 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vp7lv" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.923249 4969 scope.go:117] "RemoveContainer" containerID="9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.934806 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vp7lv"] Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.942247 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vp7lv"] Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.964151 4969 scope.go:117] "RemoveContainer" containerID="4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.990200 4969 scope.go:117] "RemoveContainer" containerID="7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f" Oct 04 09:02:38 crc kubenswrapper[4969]: E1004 09:02:38.990577 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f\": container with ID starting with 7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f not found: ID does not exist" containerID="7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.990699 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f"} err="failed to get container status \"7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f\": rpc error: code = NotFound desc = could not find container \"7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f\": container with ID starting with 7c0cd16538608a90ecb4955e392f0fed4874a22eef0852b929f974bb89d7195f not found: ID does not exist" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.990790 4969 scope.go:117] "RemoveContainer" containerID="9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9" Oct 04 09:02:38 crc kubenswrapper[4969]: E1004 09:02:38.991176 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9\": container with ID starting with 9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9 not found: ID does not exist" containerID="9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.991213 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9"} err="failed to get container status \"9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9\": rpc error: code = NotFound desc = could not find container \"9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9\": container with ID starting with 9b2ed94711808d7c026dbb9348551427ead60fa4b904da824869b3d0f1b87cf9 not found: ID does not exist" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.991235 4969 scope.go:117] "RemoveContainer" containerID="4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96" Oct 04 09:02:38 crc kubenswrapper[4969]: E1004 09:02:38.991432 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96\": container with ID starting with 4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96 not found: ID does not exist" containerID="4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96" Oct 04 09:02:38 crc kubenswrapper[4969]: I1004 09:02:38.991459 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96"} err="failed to get container status \"4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96\": rpc error: code = NotFound desc = could not find container \"4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96\": container with ID starting with 4d54aefb6bee5901cbfcee0d9c24334fa238eb59be2a76cf091051c7fd0c3d96 not found: ID does not exist" Oct 04 09:02:39 crc kubenswrapper[4969]: I1004 09:02:39.065972 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="824e57b5-550c-4c24-b243-b08c20be4a40" path="/var/lib/kubelet/pods/824e57b5-550c-4c24-b243-b08c20be4a40/volumes" Oct 04 09:02:43 crc kubenswrapper[4969]: I1004 09:02:43.969228 4969 generic.go:334] "Generic (PLEG): container finished" podID="4ce3e260-2be8-435b-8a24-7d06614c0844" containerID="2b40de07d62d74c6db7a19622db61a642550492fd4c51d91fea5a53fbc768b67" exitCode=0 Oct 04 09:02:43 crc kubenswrapper[4969]: I1004 09:02:43.969327 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" event={"ID":"4ce3e260-2be8-435b-8a24-7d06614c0844","Type":"ContainerDied","Data":"2b40de07d62d74c6db7a19622db61a642550492fd4c51d91fea5a53fbc768b67"} Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.608759 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.639358 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-1\") pod \"4ce3e260-2be8-435b-8a24-7d06614c0844\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.639479 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-extra-config-0\") pod \"4ce3e260-2be8-435b-8a24-7d06614c0844\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.639531 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-inventory\") pod \"4ce3e260-2be8-435b-8a24-7d06614c0844\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.639571 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-1\") pod \"4ce3e260-2be8-435b-8a24-7d06614c0844\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.639618 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5p5h\" (UniqueName: \"kubernetes.io/projected/4ce3e260-2be8-435b-8a24-7d06614c0844-kube-api-access-m5p5h\") pod \"4ce3e260-2be8-435b-8a24-7d06614c0844\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.639703 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-ssh-key\") pod \"4ce3e260-2be8-435b-8a24-7d06614c0844\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.639755 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-combined-ca-bundle\") pod \"4ce3e260-2be8-435b-8a24-7d06614c0844\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.639894 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-0\") pod \"4ce3e260-2be8-435b-8a24-7d06614c0844\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.639926 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-0\") pod \"4ce3e260-2be8-435b-8a24-7d06614c0844\" (UID: \"4ce3e260-2be8-435b-8a24-7d06614c0844\") " Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.677739 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4ce3e260-2be8-435b-8a24-7d06614c0844" (UID: "4ce3e260-2be8-435b-8a24-7d06614c0844"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.690878 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ce3e260-2be8-435b-8a24-7d06614c0844-kube-api-access-m5p5h" (OuterVolumeSpecName: "kube-api-access-m5p5h") pod "4ce3e260-2be8-435b-8a24-7d06614c0844" (UID: "4ce3e260-2be8-435b-8a24-7d06614c0844"). InnerVolumeSpecName "kube-api-access-m5p5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.705668 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4ce3e260-2be8-435b-8a24-7d06614c0844" (UID: "4ce3e260-2be8-435b-8a24-7d06614c0844"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.711207 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4ce3e260-2be8-435b-8a24-7d06614c0844" (UID: "4ce3e260-2be8-435b-8a24-7d06614c0844"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.712438 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-inventory" (OuterVolumeSpecName: "inventory") pod "4ce3e260-2be8-435b-8a24-7d06614c0844" (UID: "4ce3e260-2be8-435b-8a24-7d06614c0844"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.721901 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4ce3e260-2be8-435b-8a24-7d06614c0844" (UID: "4ce3e260-2be8-435b-8a24-7d06614c0844"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.727348 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4ce3e260-2be8-435b-8a24-7d06614c0844" (UID: "4ce3e260-2be8-435b-8a24-7d06614c0844"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.728466 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4ce3e260-2be8-435b-8a24-7d06614c0844" (UID: "4ce3e260-2be8-435b-8a24-7d06614c0844"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.730967 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4ce3e260-2be8-435b-8a24-7d06614c0844" (UID: "4ce3e260-2be8-435b-8a24-7d06614c0844"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.742591 4969 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.742645 4969 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.742662 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.742678 4969 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.742693 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5p5h\" (UniqueName: \"kubernetes.io/projected/4ce3e260-2be8-435b-8a24-7d06614c0844-kube-api-access-m5p5h\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.742708 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.742725 4969 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.742741 4969 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.742758 4969 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ce3e260-2be8-435b-8a24-7d06614c0844-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.998484 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" event={"ID":"4ce3e260-2be8-435b-8a24-7d06614c0844","Type":"ContainerDied","Data":"145235d3a6a698a650cfafdcb75f2d0ee1cead629ab389cca749dede1a945c40"} Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.998528 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="145235d3a6a698a650cfafdcb75f2d0ee1cead629ab389cca749dede1a945c40" Oct 04 09:02:45 crc kubenswrapper[4969]: I1004 09:02:45.998580 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p9v4f" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.234158 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp"] Oct 04 09:02:46 crc kubenswrapper[4969]: E1004 09:02:46.235551 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce3e260-2be8-435b-8a24-7d06614c0844" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.235945 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce3e260-2be8-435b-8a24-7d06614c0844" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 09:02:46 crc kubenswrapper[4969]: E1004 09:02:46.235995 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824e57b5-550c-4c24-b243-b08c20be4a40" containerName="extract-content" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.236021 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="824e57b5-550c-4c24-b243-b08c20be4a40" containerName="extract-content" Oct 04 09:02:46 crc kubenswrapper[4969]: E1004 09:02:46.236127 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824e57b5-550c-4c24-b243-b08c20be4a40" containerName="registry-server" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.236138 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="824e57b5-550c-4c24-b243-b08c20be4a40" containerName="registry-server" Oct 04 09:02:46 crc kubenswrapper[4969]: E1004 09:02:46.236175 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824e57b5-550c-4c24-b243-b08c20be4a40" containerName="extract-utilities" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.236184 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="824e57b5-550c-4c24-b243-b08c20be4a40" containerName="extract-utilities" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.238379 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce3e260-2be8-435b-8a24-7d06614c0844" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.238586 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="824e57b5-550c-4c24-b243-b08c20be4a40" containerName="registry-server" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.240117 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.243991 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.244028 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.244231 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.244874 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.245280 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gcjl8" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.255228 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp"] Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.364753 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.365031 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.365263 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.365435 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.365575 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.365708 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.365871 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z7vx\" (UniqueName: \"kubernetes.io/projected/eefed526-feaa-4158-94be-19220b2801e2-kube-api-access-6z7vx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.467533 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.467597 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.467642 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.467689 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.467782 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z7vx\" (UniqueName: \"kubernetes.io/projected/eefed526-feaa-4158-94be-19220b2801e2-kube-api-access-6z7vx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.467899 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.467937 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.473045 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.473077 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.473596 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.475398 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.478769 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.483046 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.495261 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z7vx\" (UniqueName: \"kubernetes.io/projected/eefed526-feaa-4158-94be-19220b2801e2-kube-api-access-6z7vx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k45fp\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:46 crc kubenswrapper[4969]: I1004 09:02:46.574927 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:02:47 crc kubenswrapper[4969]: I1004 09:02:47.194895 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp"] Oct 04 09:02:48 crc kubenswrapper[4969]: I1004 09:02:48.024218 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" event={"ID":"eefed526-feaa-4158-94be-19220b2801e2","Type":"ContainerStarted","Data":"fc290031f0a522880fc27ff5cfdca65e462d110b763f001057665ad0e87543ec"} Oct 04 09:02:48 crc kubenswrapper[4969]: I1004 09:02:48.024813 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" event={"ID":"eefed526-feaa-4158-94be-19220b2801e2","Type":"ContainerStarted","Data":"6cf9486dd96775d27af420f5c21de5254c24d5baa3cfd8b23d6a7441c9195df0"} Oct 04 09:02:48 crc kubenswrapper[4969]: I1004 09:02:48.054464 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" podStartSLOduration=1.556934469 podStartE2EDuration="2.054397137s" podCreationTimestamp="2025-10-04 09:02:46 +0000 UTC" firstStartedPulling="2025-10-04 09:02:47.189715639 +0000 UTC m=+2794.943984493" lastFinishedPulling="2025-10-04 09:02:47.687178337 +0000 UTC m=+2795.441447161" observedRunningTime="2025-10-04 09:02:48.045456176 +0000 UTC m=+2795.799725020" watchObservedRunningTime="2025-10-04 09:02:48.054397137 +0000 UTC m=+2795.808665991" Oct 04 09:02:49 crc kubenswrapper[4969]: I1004 09:02:49.666574 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:02:49 crc kubenswrapper[4969]: I1004 09:02:49.667099 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:03:19 crc kubenswrapper[4969]: I1004 09:03:19.666769 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:03:19 crc kubenswrapper[4969]: I1004 09:03:19.667329 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:03:19 crc kubenswrapper[4969]: I1004 09:03:19.667382 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:03:19 crc kubenswrapper[4969]: I1004 09:03:19.668255 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1de80285c1f273504a5d16302235acfdc134e2fc6fcbcc32f7101a1fbb2723c5"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:03:19 crc kubenswrapper[4969]: I1004 09:03:19.668332 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://1de80285c1f273504a5d16302235acfdc134e2fc6fcbcc32f7101a1fbb2723c5" gracePeriod=600 Oct 04 09:03:20 crc kubenswrapper[4969]: I1004 09:03:20.420346 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="1de80285c1f273504a5d16302235acfdc134e2fc6fcbcc32f7101a1fbb2723c5" exitCode=0 Oct 04 09:03:20 crc kubenswrapper[4969]: I1004 09:03:20.420451 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"1de80285c1f273504a5d16302235acfdc134e2fc6fcbcc32f7101a1fbb2723c5"} Oct 04 09:03:20 crc kubenswrapper[4969]: I1004 09:03:20.420837 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2"} Oct 04 09:03:20 crc kubenswrapper[4969]: I1004 09:03:20.420876 4969 scope.go:117] "RemoveContainer" containerID="b91bbd737f488c7492aae60c33a5a413c0c57c82cb427e4c454706e03eaf7916" Oct 04 09:05:25 crc kubenswrapper[4969]: I1004 09:05:25.942681 4969 generic.go:334] "Generic (PLEG): container finished" podID="eefed526-feaa-4158-94be-19220b2801e2" containerID="fc290031f0a522880fc27ff5cfdca65e462d110b763f001057665ad0e87543ec" exitCode=0 Oct 04 09:05:25 crc kubenswrapper[4969]: I1004 09:05:25.942929 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" event={"ID":"eefed526-feaa-4158-94be-19220b2801e2","Type":"ContainerDied","Data":"fc290031f0a522880fc27ff5cfdca65e462d110b763f001057665ad0e87543ec"} Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.428987 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.535239 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-1\") pod \"eefed526-feaa-4158-94be-19220b2801e2\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.535344 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z7vx\" (UniqueName: \"kubernetes.io/projected/eefed526-feaa-4158-94be-19220b2801e2-kube-api-access-6z7vx\") pod \"eefed526-feaa-4158-94be-19220b2801e2\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.535379 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-2\") pod \"eefed526-feaa-4158-94be-19220b2801e2\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.535485 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-inventory\") pod \"eefed526-feaa-4158-94be-19220b2801e2\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.535523 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ssh-key\") pod \"eefed526-feaa-4158-94be-19220b2801e2\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.535555 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-telemetry-combined-ca-bundle\") pod \"eefed526-feaa-4158-94be-19220b2801e2\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.535635 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-0\") pod \"eefed526-feaa-4158-94be-19220b2801e2\" (UID: \"eefed526-feaa-4158-94be-19220b2801e2\") " Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.540979 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "eefed526-feaa-4158-94be-19220b2801e2" (UID: "eefed526-feaa-4158-94be-19220b2801e2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.541391 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eefed526-feaa-4158-94be-19220b2801e2-kube-api-access-6z7vx" (OuterVolumeSpecName: "kube-api-access-6z7vx") pod "eefed526-feaa-4158-94be-19220b2801e2" (UID: "eefed526-feaa-4158-94be-19220b2801e2"). InnerVolumeSpecName "kube-api-access-6z7vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.567364 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eefed526-feaa-4158-94be-19220b2801e2" (UID: "eefed526-feaa-4158-94be-19220b2801e2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.569070 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "eefed526-feaa-4158-94be-19220b2801e2" (UID: "eefed526-feaa-4158-94be-19220b2801e2"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.569521 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "eefed526-feaa-4158-94be-19220b2801e2" (UID: "eefed526-feaa-4158-94be-19220b2801e2"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.599513 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "eefed526-feaa-4158-94be-19220b2801e2" (UID: "eefed526-feaa-4158-94be-19220b2801e2"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.600892 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-inventory" (OuterVolumeSpecName: "inventory") pod "eefed526-feaa-4158-94be-19220b2801e2" (UID: "eefed526-feaa-4158-94be-19220b2801e2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.639273 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.639338 4969 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.639371 4969 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.639400 4969 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.639449 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z7vx\" (UniqueName: \"kubernetes.io/projected/eefed526-feaa-4158-94be-19220b2801e2-kube-api-access-6z7vx\") on node \"crc\" DevicePath \"\"" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.639468 4969 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.639486 4969 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eefed526-feaa-4158-94be-19220b2801e2-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.973750 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" event={"ID":"eefed526-feaa-4158-94be-19220b2801e2","Type":"ContainerDied","Data":"6cf9486dd96775d27af420f5c21de5254c24d5baa3cfd8b23d6a7441c9195df0"} Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.974079 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cf9486dd96775d27af420f5c21de5254c24d5baa3cfd8b23d6a7441c9195df0" Oct 04 09:05:27 crc kubenswrapper[4969]: I1004 09:05:27.973814 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k45fp" Oct 04 09:05:49 crc kubenswrapper[4969]: I1004 09:05:49.668462 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:05:49 crc kubenswrapper[4969]: I1004 09:05:49.669086 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:06:03 crc kubenswrapper[4969]: I1004 09:06:03.507890 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 09:06:03 crc kubenswrapper[4969]: I1004 09:06:03.510033 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="prometheus" containerID="cri-o://7596bb8d092712e53e08b6ed3fc75137915d4fbf69e7f3231d3107c311f09e9b" gracePeriod=600 Oct 04 09:06:03 crc kubenswrapper[4969]: I1004 09:06:03.510301 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="config-reloader" containerID="cri-o://fbb54ba479493c3fb0ddbfd2ac2e97371dcae3e39fb3de63e6e2dc2f26e1f759" gracePeriod=600 Oct 04 09:06:03 crc kubenswrapper[4969]: I1004 09:06:03.510320 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="thanos-sidecar" containerID="cri-o://5f25df52f533494ca9e9131af20ae1e34564c497a31d70b745543bf8fc89fb56" gracePeriod=600 Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.056830 4969 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.137:9090/-/ready\": dial tcp 10.217.0.137:9090: connect: connection refused" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.415338 4969 generic.go:334] "Generic (PLEG): container finished" podID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerID="5f25df52f533494ca9e9131af20ae1e34564c497a31d70b745543bf8fc89fb56" exitCode=0 Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.415374 4969 generic.go:334] "Generic (PLEG): container finished" podID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerID="fbb54ba479493c3fb0ddbfd2ac2e97371dcae3e39fb3de63e6e2dc2f26e1f759" exitCode=0 Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.415387 4969 generic.go:334] "Generic (PLEG): container finished" podID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerID="7596bb8d092712e53e08b6ed3fc75137915d4fbf69e7f3231d3107c311f09e9b" exitCode=0 Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.415409 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerDied","Data":"5f25df52f533494ca9e9131af20ae1e34564c497a31d70b745543bf8fc89fb56"} Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.415491 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerDied","Data":"fbb54ba479493c3fb0ddbfd2ac2e97371dcae3e39fb3de63e6e2dc2f26e1f759"} Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.415504 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerDied","Data":"7596bb8d092712e53e08b6ed3fc75137915d4fbf69e7f3231d3107c311f09e9b"} Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.552123 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.726839 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.726901 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.727103 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.727125 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-secret-combined-ca-bundle\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.727181 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.727227 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.727271 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-thanos-prometheus-http-client-file\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.727309 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-tls-assets\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.727355 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9d35bef4-01ad-4c3d-b042-9ac2350420a4-prometheus-metric-storage-rulefiles-0\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.727405 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config-out\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.727472 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqxww\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-kube-api-access-tqxww\") pod \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\" (UID: \"9d35bef4-01ad-4c3d-b042-9ac2350420a4\") " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.728024 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d35bef4-01ad-4c3d-b042-9ac2350420a4-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.734862 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config-out" (OuterVolumeSpecName: "config-out") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.753901 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.753918 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.753999 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.754072 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.754122 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config" (OuterVolumeSpecName: "config") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.754186 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.762299 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-kube-api-access-tqxww" (OuterVolumeSpecName: "kube-api-access-tqxww") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "kube-api-access-tqxww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.774302 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.829140 4969 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") on node \"crc\" " Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.835564 4969 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.835811 4969 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.835899 4969 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.835990 4969 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.836072 4969 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9d35bef4-01ad-4c3d-b042-9ac2350420a4-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.836164 4969 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9d35bef4-01ad-4c3d-b042-9ac2350420a4-config-out\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.836253 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqxww\" (UniqueName: \"kubernetes.io/projected/9d35bef4-01ad-4c3d-b042-9ac2350420a4-kube-api-access-tqxww\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.836607 4969 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.836691 4969 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.844326 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config" (OuterVolumeSpecName: "web-config") pod "9d35bef4-01ad-4c3d-b042-9ac2350420a4" (UID: "9d35bef4-01ad-4c3d-b042-9ac2350420a4"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.860066 4969 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.860927 4969 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a") on node "crc" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.938038 4969 reconciler_common.go:293] "Volume detached for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:04 crc kubenswrapper[4969]: I1004 09:06:04.938074 4969 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9d35bef4-01ad-4c3d-b042-9ac2350420a4-web-config\") on node \"crc\" DevicePath \"\"" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.431884 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9d35bef4-01ad-4c3d-b042-9ac2350420a4","Type":"ContainerDied","Data":"4f3c57ab7e7f4de9e58e0516b0c3ec637a1a71eb9a032938c141cf03b1ce54c5"} Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.431968 4969 scope.go:117] "RemoveContainer" containerID="5f25df52f533494ca9e9131af20ae1e34564c497a31d70b745543bf8fc89fb56" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.431986 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.463853 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.480204 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.493124 4969 scope.go:117] "RemoveContainer" containerID="fbb54ba479493c3fb0ddbfd2ac2e97371dcae3e39fb3de63e6e2dc2f26e1f759" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.497599 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 09:06:05 crc kubenswrapper[4969]: E1004 09:06:05.498019 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="thanos-sidecar" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.498041 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="thanos-sidecar" Oct 04 09:06:05 crc kubenswrapper[4969]: E1004 09:06:05.498063 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="prometheus" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.498074 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="prometheus" Oct 04 09:06:05 crc kubenswrapper[4969]: E1004 09:06:05.498110 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="init-config-reloader" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.498119 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="init-config-reloader" Oct 04 09:06:05 crc kubenswrapper[4969]: E1004 09:06:05.498148 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="config-reloader" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.498157 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="config-reloader" Oct 04 09:06:05 crc kubenswrapper[4969]: E1004 09:06:05.498179 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eefed526-feaa-4158-94be-19220b2801e2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.498192 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="eefed526-feaa-4158-94be-19220b2801e2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.498505 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="prometheus" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.498552 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="config-reloader" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.498573 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" containerName="thanos-sidecar" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.498588 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="eefed526-feaa-4158-94be-19220b2801e2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.501096 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.507064 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.512602 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.516232 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.516605 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-n8xt6" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.516746 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.524558 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.532739 4969 scope.go:117] "RemoveContainer" containerID="7596bb8d092712e53e08b6ed3fc75137915d4fbf69e7f3231d3107c311f09e9b" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.550510 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.573886 4969 scope.go:117] "RemoveContainer" containerID="5ae2543215bb1ced07079e957d140539dd1ef5d1a783c89a8c8fe25e15fca67b" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.662358 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.662434 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.662458 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.662480 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc7rl\" (UniqueName: \"kubernetes.io/projected/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-kube-api-access-rc7rl\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.662510 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.662547 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.662599 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-config\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.663936 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.663990 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.664010 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.664028 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765315 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765358 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765378 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765401 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765446 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765479 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765499 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765525 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc7rl\" (UniqueName: \"kubernetes.io/projected/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-kube-api-access-rc7rl\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765550 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765587 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.765637 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-config\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.766312 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.771183 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.772218 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.772265 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.772729 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.774158 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-config\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.774936 4969 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.774997 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6bc02e5d9caae1a8b7f65724c18b281ae808553b7b037139e0479e345eca79b6/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.775332 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.782124 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.784785 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.785026 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc7rl\" (UniqueName: \"kubernetes.io/projected/9c93fd81-fe7a-4e9f-83ef-166655d4d8d0-kube-api-access-rc7rl\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.829599 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20938cd1-ee35-4a12-b38a-fdd63743d86a\") pod \"prometheus-metric-storage-0\" (UID: \"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0\") " pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:05 crc kubenswrapper[4969]: I1004 09:06:05.839971 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:06 crc kubenswrapper[4969]: I1004 09:06:06.342730 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 09:06:06 crc kubenswrapper[4969]: I1004 09:06:06.461846 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0","Type":"ContainerStarted","Data":"477dc813df6f81007726c1e0742861dc0e8647f2de2665d2f2978641b3bd2b50"} Oct 04 09:06:07 crc kubenswrapper[4969]: I1004 09:06:07.071874 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d35bef4-01ad-4c3d-b042-9ac2350420a4" path="/var/lib/kubelet/pods/9d35bef4-01ad-4c3d-b042-9ac2350420a4/volumes" Oct 04 09:06:11 crc kubenswrapper[4969]: I1004 09:06:11.523365 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0","Type":"ContainerStarted","Data":"8bb5726dd401dbd3620bf19de09158a9dfd0e4fa2a1c393f9d2994ac1254792f"} Oct 04 09:06:19 crc kubenswrapper[4969]: I1004 09:06:19.667188 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:06:19 crc kubenswrapper[4969]: I1004 09:06:19.668141 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:06:22 crc kubenswrapper[4969]: I1004 09:06:22.686146 4969 generic.go:334] "Generic (PLEG): container finished" podID="9c93fd81-fe7a-4e9f-83ef-166655d4d8d0" containerID="8bb5726dd401dbd3620bf19de09158a9dfd0e4fa2a1c393f9d2994ac1254792f" exitCode=0 Oct 04 09:06:22 crc kubenswrapper[4969]: I1004 09:06:22.688584 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0","Type":"ContainerDied","Data":"8bb5726dd401dbd3620bf19de09158a9dfd0e4fa2a1c393f9d2994ac1254792f"} Oct 04 09:06:23 crc kubenswrapper[4969]: I1004 09:06:23.703728 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0","Type":"ContainerStarted","Data":"628904a6b09b3bf748dd616e344a229af0b0e1da085788b386f1519a6abf067c"} Oct 04 09:06:27 crc kubenswrapper[4969]: I1004 09:06:27.749600 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0","Type":"ContainerStarted","Data":"56b49e799a620b607caf4da848e6c08dd525f5bf4d01aa967bc2901570fb31ed"} Oct 04 09:06:28 crc kubenswrapper[4969]: I1004 09:06:28.771348 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9c93fd81-fe7a-4e9f-83ef-166655d4d8d0","Type":"ContainerStarted","Data":"8ee89355c255ff65a16bbb9b8400aba145e1c0a2f2970b5a8ae36b0a1661922d"} Oct 04 09:06:28 crc kubenswrapper[4969]: I1004 09:06:28.826844 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=23.826817097 podStartE2EDuration="23.826817097s" podCreationTimestamp="2025-10-04 09:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 09:06:28.81398856 +0000 UTC m=+3016.568257454" watchObservedRunningTime="2025-10-04 09:06:28.826817097 +0000 UTC m=+3016.581085951" Oct 04 09:06:30 crc kubenswrapper[4969]: I1004 09:06:30.840936 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:35 crc kubenswrapper[4969]: I1004 09:06:35.843854 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:35 crc kubenswrapper[4969]: I1004 09:06:35.851756 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:35 crc kubenswrapper[4969]: I1004 09:06:35.896994 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 04 09:06:49 crc kubenswrapper[4969]: I1004 09:06:49.667222 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:06:49 crc kubenswrapper[4969]: I1004 09:06:49.667864 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:06:49 crc kubenswrapper[4969]: I1004 09:06:49.667920 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:06:49 crc kubenswrapper[4969]: I1004 09:06:49.668783 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:06:49 crc kubenswrapper[4969]: I1004 09:06:49.668851 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" gracePeriod=600 Oct 04 09:06:49 crc kubenswrapper[4969]: E1004 09:06:49.795271 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:06:50 crc kubenswrapper[4969]: I1004 09:06:50.061684 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" exitCode=0 Oct 04 09:06:50 crc kubenswrapper[4969]: I1004 09:06:50.061753 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2"} Oct 04 09:06:50 crc kubenswrapper[4969]: I1004 09:06:50.061802 4969 scope.go:117] "RemoveContainer" containerID="1de80285c1f273504a5d16302235acfdc134e2fc6fcbcc32f7101a1fbb2723c5" Oct 04 09:06:50 crc kubenswrapper[4969]: I1004 09:06:50.062921 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:06:50 crc kubenswrapper[4969]: E1004 09:06:50.063510 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.058142 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:07:01 crc kubenswrapper[4969]: E1004 09:07:01.059198 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.649366 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.650595 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.655501 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.657840 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.658038 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.658467 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gh4r4" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.676252 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.766079 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.766335 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.766741 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-config-data\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.766812 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.766920 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.766965 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.767189 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.767239 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqw2t\" (UniqueName: \"kubernetes.io/projected/6348f782-6e3b-4c67-b252-2276628c14d6-kube-api-access-qqw2t\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.767542 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869338 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869405 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqw2t\" (UniqueName: \"kubernetes.io/projected/6348f782-6e3b-4c67-b252-2276628c14d6-kube-api-access-qqw2t\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869499 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869546 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869623 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869822 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-config-data\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869863 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869899 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869929 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.869963 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.870990 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.871762 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-config-data\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.872040 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.872848 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.878124 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.878634 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.881600 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.896393 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqw2t\" (UniqueName: \"kubernetes.io/projected/6348f782-6e3b-4c67-b252-2276628c14d6-kube-api-access-qqw2t\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.922414 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " pod="openstack/tempest-tests-tempest" Oct 04 09:07:01 crc kubenswrapper[4969]: I1004 09:07:01.982788 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 09:07:02 crc kubenswrapper[4969]: I1004 09:07:02.461830 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 09:07:02 crc kubenswrapper[4969]: I1004 09:07:02.467795 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 09:07:03 crc kubenswrapper[4969]: I1004 09:07:03.239393 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"6348f782-6e3b-4c67-b252-2276628c14d6","Type":"ContainerStarted","Data":"ac15602ac98165544941c206deea5027ba468ded13f8a6c01ffc5b7dae08684a"} Oct 04 09:07:14 crc kubenswrapper[4969]: I1004 09:07:14.055563 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:07:14 crc kubenswrapper[4969]: E1004 09:07:14.056666 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:07:24 crc kubenswrapper[4969]: E1004 09:07:24.313815 4969 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-tempest-all:watcher_latest" Oct 04 09:07:24 crc kubenswrapper[4969]: E1004 09:07:24.314495 4969 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.83:5001/podified-master-centos10/openstack-tempest-all:watcher_latest" Oct 04 09:07:24 crc kubenswrapper[4969]: E1004 09:07:24.314750 4969 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:38.102.83.83:5001/podified-master-centos10/openstack-tempest-all:watcher_latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qqw2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(6348f782-6e3b-4c67-b252-2276628c14d6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 09:07:24 crc kubenswrapper[4969]: E1004 09:07:24.316080 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="6348f782-6e3b-4c67-b252-2276628c14d6" Oct 04 09:07:24 crc kubenswrapper[4969]: E1004 09:07:24.516246 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.83:5001/podified-master-centos10/openstack-tempest-all:watcher_latest\\\"\"" pod="openstack/tempest-tests-tempest" podUID="6348f782-6e3b-4c67-b252-2276628c14d6" Oct 04 09:07:27 crc kubenswrapper[4969]: I1004 09:07:27.056814 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:07:27 crc kubenswrapper[4969]: E1004 09:07:27.057577 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:07:38 crc kubenswrapper[4969]: I1004 09:07:38.123819 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 09:07:39 crc kubenswrapper[4969]: I1004 09:07:39.707396 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"6348f782-6e3b-4c67-b252-2276628c14d6","Type":"ContainerStarted","Data":"95da218138474e9dafb1c5e2f0055868b9b87f05b47cec19c4abea56f80ae0ed"} Oct 04 09:07:39 crc kubenswrapper[4969]: I1004 09:07:39.736681 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.082783651 podStartE2EDuration="39.736657282s" podCreationTimestamp="2025-10-04 09:07:00 +0000 UTC" firstStartedPulling="2025-10-04 09:07:02.467621245 +0000 UTC m=+3050.221890059" lastFinishedPulling="2025-10-04 09:07:38.121494856 +0000 UTC m=+3085.875763690" observedRunningTime="2025-10-04 09:07:39.729623488 +0000 UTC m=+3087.483892332" watchObservedRunningTime="2025-10-04 09:07:39.736657282 +0000 UTC m=+3087.490926116" Oct 04 09:07:40 crc kubenswrapper[4969]: I1004 09:07:40.055485 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:07:40 crc kubenswrapper[4969]: E1004 09:07:40.055938 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:07:52 crc kubenswrapper[4969]: I1004 09:07:52.055510 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:07:52 crc kubenswrapper[4969]: E1004 09:07:52.056588 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:08:04 crc kubenswrapper[4969]: I1004 09:08:04.055414 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:08:04 crc kubenswrapper[4969]: E1004 09:08:04.056518 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:08:18 crc kubenswrapper[4969]: I1004 09:08:18.055567 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:08:18 crc kubenswrapper[4969]: E1004 09:08:18.056645 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:08:30 crc kubenswrapper[4969]: I1004 09:08:30.055112 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:08:30 crc kubenswrapper[4969]: E1004 09:08:30.055923 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:08:43 crc kubenswrapper[4969]: I1004 09:08:43.061623 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:08:43 crc kubenswrapper[4969]: E1004 09:08:43.062333 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:08:57 crc kubenswrapper[4969]: I1004 09:08:57.055924 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:08:57 crc kubenswrapper[4969]: E1004 09:08:57.057012 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:09:09 crc kubenswrapper[4969]: I1004 09:09:09.056604 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:09:09 crc kubenswrapper[4969]: E1004 09:09:09.057609 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:09:24 crc kubenswrapper[4969]: I1004 09:09:24.055187 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:09:24 crc kubenswrapper[4969]: E1004 09:09:24.058161 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:09:36 crc kubenswrapper[4969]: I1004 09:09:36.056698 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:09:36 crc kubenswrapper[4969]: E1004 09:09:36.057821 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:09:48 crc kubenswrapper[4969]: I1004 09:09:48.056319 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:09:48 crc kubenswrapper[4969]: E1004 09:09:48.057490 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:10:03 crc kubenswrapper[4969]: I1004 09:10:03.094922 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:10:03 crc kubenswrapper[4969]: E1004 09:10:03.095901 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.640016 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dwp8w"] Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.643049 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.650010 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dwp8w"] Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.695295 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdlng\" (UniqueName: \"kubernetes.io/projected/b03f34a4-f440-42a4-8496-5590681e9e90-kube-api-access-jdlng\") pod \"certified-operators-dwp8w\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.695473 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-catalog-content\") pod \"certified-operators-dwp8w\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.695711 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-utilities\") pod \"certified-operators-dwp8w\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.797640 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-catalog-content\") pod \"certified-operators-dwp8w\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.797947 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-utilities\") pod \"certified-operators-dwp8w\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.798071 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-catalog-content\") pod \"certified-operators-dwp8w\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.798335 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-utilities\") pod \"certified-operators-dwp8w\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.798368 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdlng\" (UniqueName: \"kubernetes.io/projected/b03f34a4-f440-42a4-8496-5590681e9e90-kube-api-access-jdlng\") pod \"certified-operators-dwp8w\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.816714 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdlng\" (UniqueName: \"kubernetes.io/projected/b03f34a4-f440-42a4-8496-5590681e9e90-kube-api-access-jdlng\") pod \"certified-operators-dwp8w\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:05 crc kubenswrapper[4969]: I1004 09:10:05.964208 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:06 crc kubenswrapper[4969]: I1004 09:10:06.448544 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dwp8w"] Oct 04 09:10:06 crc kubenswrapper[4969]: I1004 09:10:06.620315 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dwp8w" event={"ID":"b03f34a4-f440-42a4-8496-5590681e9e90","Type":"ContainerStarted","Data":"14927852c063fdd4e8ea3fa52c8c40cd88f8b5633d9d7c78a2499bd913fcd3fd"} Oct 04 09:10:07 crc kubenswrapper[4969]: I1004 09:10:07.634497 4969 generic.go:334] "Generic (PLEG): container finished" podID="b03f34a4-f440-42a4-8496-5590681e9e90" containerID="66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b" exitCode=0 Oct 04 09:10:07 crc kubenswrapper[4969]: I1004 09:10:07.634577 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dwp8w" event={"ID":"b03f34a4-f440-42a4-8496-5590681e9e90","Type":"ContainerDied","Data":"66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b"} Oct 04 09:10:09 crc kubenswrapper[4969]: I1004 09:10:09.662641 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dwp8w" event={"ID":"b03f34a4-f440-42a4-8496-5590681e9e90","Type":"ContainerStarted","Data":"c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322"} Oct 04 09:10:18 crc kubenswrapper[4969]: I1004 09:10:18.055800 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:10:18 crc kubenswrapper[4969]: E1004 09:10:18.057097 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:10:18 crc kubenswrapper[4969]: I1004 09:10:18.747793 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a525e7b-7a1c-4671-a6cc-e81bd316fd48" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Oct 04 09:10:23 crc kubenswrapper[4969]: I1004 09:10:23.751981 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a525e7b-7a1c-4671-a6cc-e81bd316fd48" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Oct 04 09:10:23 crc kubenswrapper[4969]: I1004 09:10:23.752309 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a525e7b-7a1c-4671-a6cc-e81bd316fd48" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Oct 04 09:10:24 crc kubenswrapper[4969]: I1004 09:10:24.135760 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a525e7b-7a1c-4671-a6cc-e81bd316fd48" containerName="ceilometer-central-agent" probeResult="failure" output=< Oct 04 09:10:24 crc kubenswrapper[4969]: Unkown error: Expecting value: line 1 column 1 (char 0) Oct 04 09:10:24 crc kubenswrapper[4969]: > Oct 04 09:10:24 crc kubenswrapper[4969]: I1004 09:10:24.135841 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Oct 04 09:10:24 crc kubenswrapper[4969]: I1004 09:10:24.136638 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"55844fb58089e4b9ffac573acfa005875b27c7ae60162968ad9c061714b8d232"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Oct 04 09:10:24 crc kubenswrapper[4969]: I1004 09:10:24.136715 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a525e7b-7a1c-4671-a6cc-e81bd316fd48" containerName="ceilometer-central-agent" containerID="cri-o://55844fb58089e4b9ffac573acfa005875b27c7ae60162968ad9c061714b8d232" gracePeriod=30 Oct 04 09:10:24 crc kubenswrapper[4969]: I1004 09:10:24.855857 4969 generic.go:334] "Generic (PLEG): container finished" podID="b03f34a4-f440-42a4-8496-5590681e9e90" containerID="c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322" exitCode=0 Oct 04 09:10:24 crc kubenswrapper[4969]: I1004 09:10:24.855922 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dwp8w" event={"ID":"b03f34a4-f440-42a4-8496-5590681e9e90","Type":"ContainerDied","Data":"c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322"} Oct 04 09:10:25 crc kubenswrapper[4969]: I1004 09:10:25.872689 4969 generic.go:334] "Generic (PLEG): container finished" podID="0a525e7b-7a1c-4671-a6cc-e81bd316fd48" containerID="55844fb58089e4b9ffac573acfa005875b27c7ae60162968ad9c061714b8d232" exitCode=0 Oct 04 09:10:25 crc kubenswrapper[4969]: I1004 09:10:25.872761 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a525e7b-7a1c-4671-a6cc-e81bd316fd48","Type":"ContainerDied","Data":"55844fb58089e4b9ffac573acfa005875b27c7ae60162968ad9c061714b8d232"} Oct 04 09:10:28 crc kubenswrapper[4969]: I1004 09:10:28.918223 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a525e7b-7a1c-4671-a6cc-e81bd316fd48","Type":"ContainerStarted","Data":"3d378e82596f3ef000ef6e925b8f96a18d335677d836a64705d5deeec3def5b2"} Oct 04 09:10:28 crc kubenswrapper[4969]: I1004 09:10:28.921871 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dwp8w" event={"ID":"b03f34a4-f440-42a4-8496-5590681e9e90","Type":"ContainerStarted","Data":"786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4"} Oct 04 09:10:28 crc kubenswrapper[4969]: I1004 09:10:28.965952 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dwp8w" podStartSLOduration=3.231632999 podStartE2EDuration="23.965931993s" podCreationTimestamp="2025-10-04 09:10:05 +0000 UTC" firstStartedPulling="2025-10-04 09:10:07.63705 +0000 UTC m=+3235.391318824" lastFinishedPulling="2025-10-04 09:10:28.371348974 +0000 UTC m=+3256.125617818" observedRunningTime="2025-10-04 09:10:28.955276669 +0000 UTC m=+3256.709545513" watchObservedRunningTime="2025-10-04 09:10:28.965931993 +0000 UTC m=+3256.720200817" Oct 04 09:10:32 crc kubenswrapper[4969]: I1004 09:10:32.054994 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:10:32 crc kubenswrapper[4969]: E1004 09:10:32.055511 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:10:35 crc kubenswrapper[4969]: I1004 09:10:35.964493 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:35 crc kubenswrapper[4969]: I1004 09:10:35.965136 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:36 crc kubenswrapper[4969]: I1004 09:10:36.091587 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:36 crc kubenswrapper[4969]: I1004 09:10:36.144174 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:36 crc kubenswrapper[4969]: I1004 09:10:36.836844 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dwp8w"] Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.029587 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dwp8w" podUID="b03f34a4-f440-42a4-8496-5590681e9e90" containerName="registry-server" containerID="cri-o://786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4" gracePeriod=2 Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.594677 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.731592 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdlng\" (UniqueName: \"kubernetes.io/projected/b03f34a4-f440-42a4-8496-5590681e9e90-kube-api-access-jdlng\") pod \"b03f34a4-f440-42a4-8496-5590681e9e90\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.731747 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-utilities\") pod \"b03f34a4-f440-42a4-8496-5590681e9e90\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.731848 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-catalog-content\") pod \"b03f34a4-f440-42a4-8496-5590681e9e90\" (UID: \"b03f34a4-f440-42a4-8496-5590681e9e90\") " Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.734304 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-utilities" (OuterVolumeSpecName: "utilities") pod "b03f34a4-f440-42a4-8496-5590681e9e90" (UID: "b03f34a4-f440-42a4-8496-5590681e9e90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.743043 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03f34a4-f440-42a4-8496-5590681e9e90-kube-api-access-jdlng" (OuterVolumeSpecName: "kube-api-access-jdlng") pod "b03f34a4-f440-42a4-8496-5590681e9e90" (UID: "b03f34a4-f440-42a4-8496-5590681e9e90"). InnerVolumeSpecName "kube-api-access-jdlng". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.776924 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b03f34a4-f440-42a4-8496-5590681e9e90" (UID: "b03f34a4-f440-42a4-8496-5590681e9e90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.834679 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.834732 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdlng\" (UniqueName: \"kubernetes.io/projected/b03f34a4-f440-42a4-8496-5590681e9e90-kube-api-access-jdlng\") on node \"crc\" DevicePath \"\"" Oct 04 09:10:38 crc kubenswrapper[4969]: I1004 09:10:38.834754 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03f34a4-f440-42a4-8496-5590681e9e90-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.043916 4969 generic.go:334] "Generic (PLEG): container finished" podID="b03f34a4-f440-42a4-8496-5590681e9e90" containerID="786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4" exitCode=0 Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.043981 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dwp8w" event={"ID":"b03f34a4-f440-42a4-8496-5590681e9e90","Type":"ContainerDied","Data":"786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4"} Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.044020 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dwp8w" event={"ID":"b03f34a4-f440-42a4-8496-5590681e9e90","Type":"ContainerDied","Data":"14927852c063fdd4e8ea3fa52c8c40cd88f8b5633d9d7c78a2499bd913fcd3fd"} Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.044049 4969 scope.go:117] "RemoveContainer" containerID="786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.044218 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dwp8w" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.090634 4969 scope.go:117] "RemoveContainer" containerID="c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.131686 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dwp8w"] Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.143901 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dwp8w"] Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.144914 4969 scope.go:117] "RemoveContainer" containerID="66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.198842 4969 scope.go:117] "RemoveContainer" containerID="786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4" Oct 04 09:10:39 crc kubenswrapper[4969]: E1004 09:10:39.199190 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4\": container with ID starting with 786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4 not found: ID does not exist" containerID="786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.199229 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4"} err="failed to get container status \"786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4\": rpc error: code = NotFound desc = could not find container \"786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4\": container with ID starting with 786b786c5aece8058bf647ce8682e9b013d8a7d317995744db5af4ef847084f4 not found: ID does not exist" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.199253 4969 scope.go:117] "RemoveContainer" containerID="c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322" Oct 04 09:10:39 crc kubenswrapper[4969]: E1004 09:10:39.199445 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322\": container with ID starting with c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322 not found: ID does not exist" containerID="c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.199470 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322"} err="failed to get container status \"c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322\": rpc error: code = NotFound desc = could not find container \"c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322\": container with ID starting with c5ba5c71e999bce5f085460592af24ae2113b0bb82f0f8a5e6f2d9a5434e0322 not found: ID does not exist" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.199490 4969 scope.go:117] "RemoveContainer" containerID="66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b" Oct 04 09:10:39 crc kubenswrapper[4969]: E1004 09:10:39.199703 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b\": container with ID starting with 66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b not found: ID does not exist" containerID="66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b" Oct 04 09:10:39 crc kubenswrapper[4969]: I1004 09:10:39.199724 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b"} err="failed to get container status \"66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b\": rpc error: code = NotFound desc = could not find container \"66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b\": container with ID starting with 66334241398f0d8ae6bdc54420eeca8ab3979889d87a45d30260353bfdf8af3b not found: ID does not exist" Oct 04 09:10:41 crc kubenswrapper[4969]: I1004 09:10:41.082361 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03f34a4-f440-42a4-8496-5590681e9e90" path="/var/lib/kubelet/pods/b03f34a4-f440-42a4-8496-5590681e9e90/volumes" Oct 04 09:10:43 crc kubenswrapper[4969]: I1004 09:10:43.068119 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:10:43 crc kubenswrapper[4969]: E1004 09:10:43.068798 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:10:54 crc kubenswrapper[4969]: I1004 09:10:54.054498 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:10:54 crc kubenswrapper[4969]: E1004 09:10:54.055179 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:11:08 crc kubenswrapper[4969]: I1004 09:11:08.055726 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:11:08 crc kubenswrapper[4969]: E1004 09:11:08.056992 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:11:23 crc kubenswrapper[4969]: I1004 09:11:23.055475 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:11:23 crc kubenswrapper[4969]: E1004 09:11:23.056529 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:11:34 crc kubenswrapper[4969]: I1004 09:11:34.056563 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:11:34 crc kubenswrapper[4969]: E1004 09:11:34.057803 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:11:47 crc kubenswrapper[4969]: I1004 09:11:47.056528 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:11:47 crc kubenswrapper[4969]: E1004 09:11:47.057862 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.598077 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c25vm"] Oct 04 09:11:57 crc kubenswrapper[4969]: E1004 09:11:57.599198 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03f34a4-f440-42a4-8496-5590681e9e90" containerName="registry-server" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.599220 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03f34a4-f440-42a4-8496-5590681e9e90" containerName="registry-server" Oct 04 09:11:57 crc kubenswrapper[4969]: E1004 09:11:57.599246 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03f34a4-f440-42a4-8496-5590681e9e90" containerName="extract-content" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.599259 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03f34a4-f440-42a4-8496-5590681e9e90" containerName="extract-content" Oct 04 09:11:57 crc kubenswrapper[4969]: E1004 09:11:57.599296 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03f34a4-f440-42a4-8496-5590681e9e90" containerName="extract-utilities" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.599308 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03f34a4-f440-42a4-8496-5590681e9e90" containerName="extract-utilities" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.599662 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03f34a4-f440-42a4-8496-5590681e9e90" containerName="registry-server" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.602147 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.612980 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c25vm"] Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.736092 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r62t\" (UniqueName: \"kubernetes.io/projected/5e79a99d-a604-422f-9711-d8df00798f3c-kube-api-access-2r62t\") pod \"community-operators-c25vm\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.736377 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-utilities\") pod \"community-operators-c25vm\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.736440 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-catalog-content\") pod \"community-operators-c25vm\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.838598 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r62t\" (UniqueName: \"kubernetes.io/projected/5e79a99d-a604-422f-9711-d8df00798f3c-kube-api-access-2r62t\") pod \"community-operators-c25vm\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.838942 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-utilities\") pod \"community-operators-c25vm\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.839063 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-catalog-content\") pod \"community-operators-c25vm\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.839345 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-utilities\") pod \"community-operators-c25vm\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.839659 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-catalog-content\") pod \"community-operators-c25vm\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.865282 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r62t\" (UniqueName: \"kubernetes.io/projected/5e79a99d-a604-422f-9711-d8df00798f3c-kube-api-access-2r62t\") pod \"community-operators-c25vm\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:57 crc kubenswrapper[4969]: I1004 09:11:57.936787 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:11:58 crc kubenswrapper[4969]: W1004 09:11:58.434828 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e79a99d_a604_422f_9711_d8df00798f3c.slice/crio-734702c123faf63d4f144bf42413744333682090f1cf742f8b2e88be35b3c6e2 WatchSource:0}: Error finding container 734702c123faf63d4f144bf42413744333682090f1cf742f8b2e88be35b3c6e2: Status 404 returned error can't find the container with id 734702c123faf63d4f144bf42413744333682090f1cf742f8b2e88be35b3c6e2 Oct 04 09:11:58 crc kubenswrapper[4969]: I1004 09:11:58.438960 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c25vm"] Oct 04 09:11:59 crc kubenswrapper[4969]: I1004 09:11:59.025648 4969 generic.go:334] "Generic (PLEG): container finished" podID="5e79a99d-a604-422f-9711-d8df00798f3c" containerID="bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695" exitCode=0 Oct 04 09:11:59 crc kubenswrapper[4969]: I1004 09:11:59.025737 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c25vm" event={"ID":"5e79a99d-a604-422f-9711-d8df00798f3c","Type":"ContainerDied","Data":"bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695"} Oct 04 09:11:59 crc kubenswrapper[4969]: I1004 09:11:59.025999 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c25vm" event={"ID":"5e79a99d-a604-422f-9711-d8df00798f3c","Type":"ContainerStarted","Data":"734702c123faf63d4f144bf42413744333682090f1cf742f8b2e88be35b3c6e2"} Oct 04 09:11:59 crc kubenswrapper[4969]: I1004 09:11:59.056414 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:12:00 crc kubenswrapper[4969]: I1004 09:12:00.040040 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"6996838e394ee153292731315425104062d2c84f42bbc656b3b6c05ce870ab37"} Oct 04 09:12:01 crc kubenswrapper[4969]: I1004 09:12:01.052007 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c25vm" event={"ID":"5e79a99d-a604-422f-9711-d8df00798f3c","Type":"ContainerStarted","Data":"e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3"} Oct 04 09:12:04 crc kubenswrapper[4969]: I1004 09:12:04.097824 4969 generic.go:334] "Generic (PLEG): container finished" podID="5e79a99d-a604-422f-9711-d8df00798f3c" containerID="e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3" exitCode=0 Oct 04 09:12:04 crc kubenswrapper[4969]: I1004 09:12:04.097928 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c25vm" event={"ID":"5e79a99d-a604-422f-9711-d8df00798f3c","Type":"ContainerDied","Data":"e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3"} Oct 04 09:12:04 crc kubenswrapper[4969]: I1004 09:12:04.102571 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 09:12:05 crc kubenswrapper[4969]: I1004 09:12:05.110143 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c25vm" event={"ID":"5e79a99d-a604-422f-9711-d8df00798f3c","Type":"ContainerStarted","Data":"87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb"} Oct 04 09:12:05 crc kubenswrapper[4969]: I1004 09:12:05.140478 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c25vm" podStartSLOduration=2.375859605 podStartE2EDuration="8.140464856s" podCreationTimestamp="2025-10-04 09:11:57 +0000 UTC" firstStartedPulling="2025-10-04 09:11:59.028256773 +0000 UTC m=+3346.782525627" lastFinishedPulling="2025-10-04 09:12:04.792862024 +0000 UTC m=+3352.547130878" observedRunningTime="2025-10-04 09:12:05.133027361 +0000 UTC m=+3352.887296185" watchObservedRunningTime="2025-10-04 09:12:05.140464856 +0000 UTC m=+3352.894733670" Oct 04 09:12:07 crc kubenswrapper[4969]: I1004 09:12:07.937005 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:12:07 crc kubenswrapper[4969]: I1004 09:12:07.938718 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:12:09 crc kubenswrapper[4969]: I1004 09:12:09.019505 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-c25vm" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" containerName="registry-server" probeResult="failure" output=< Oct 04 09:12:09 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:12:09 crc kubenswrapper[4969]: > Oct 04 09:12:17 crc kubenswrapper[4969]: I1004 09:12:17.987039 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:12:18 crc kubenswrapper[4969]: I1004 09:12:18.047491 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:12:18 crc kubenswrapper[4969]: I1004 09:12:18.241388 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c25vm"] Oct 04 09:12:19 crc kubenswrapper[4969]: I1004 09:12:19.295606 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c25vm" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" containerName="registry-server" containerID="cri-o://87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb" gracePeriod=2 Oct 04 09:12:19 crc kubenswrapper[4969]: I1004 09:12:19.879573 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:12:19 crc kubenswrapper[4969]: I1004 09:12:19.972429 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r62t\" (UniqueName: \"kubernetes.io/projected/5e79a99d-a604-422f-9711-d8df00798f3c-kube-api-access-2r62t\") pod \"5e79a99d-a604-422f-9711-d8df00798f3c\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " Oct 04 09:12:19 crc kubenswrapper[4969]: I1004 09:12:19.972524 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-utilities\") pod \"5e79a99d-a604-422f-9711-d8df00798f3c\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " Oct 04 09:12:19 crc kubenswrapper[4969]: I1004 09:12:19.972624 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-catalog-content\") pod \"5e79a99d-a604-422f-9711-d8df00798f3c\" (UID: \"5e79a99d-a604-422f-9711-d8df00798f3c\") " Oct 04 09:12:19 crc kubenswrapper[4969]: I1004 09:12:19.973572 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-utilities" (OuterVolumeSpecName: "utilities") pod "5e79a99d-a604-422f-9711-d8df00798f3c" (UID: "5e79a99d-a604-422f-9711-d8df00798f3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:12:19 crc kubenswrapper[4969]: I1004 09:12:19.979732 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e79a99d-a604-422f-9711-d8df00798f3c-kube-api-access-2r62t" (OuterVolumeSpecName: "kube-api-access-2r62t") pod "5e79a99d-a604-422f-9711-d8df00798f3c" (UID: "5e79a99d-a604-422f-9711-d8df00798f3c"). InnerVolumeSpecName "kube-api-access-2r62t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.024678 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e79a99d-a604-422f-9711-d8df00798f3c" (UID: "5e79a99d-a604-422f-9711-d8df00798f3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.075275 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r62t\" (UniqueName: \"kubernetes.io/projected/5e79a99d-a604-422f-9711-d8df00798f3c-kube-api-access-2r62t\") on node \"crc\" DevicePath \"\"" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.075322 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.075339 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e79a99d-a604-422f-9711-d8df00798f3c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.308870 4969 generic.go:334] "Generic (PLEG): container finished" podID="5e79a99d-a604-422f-9711-d8df00798f3c" containerID="87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb" exitCode=0 Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.308934 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c25vm" event={"ID":"5e79a99d-a604-422f-9711-d8df00798f3c","Type":"ContainerDied","Data":"87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb"} Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.308975 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c25vm" event={"ID":"5e79a99d-a604-422f-9711-d8df00798f3c","Type":"ContainerDied","Data":"734702c123faf63d4f144bf42413744333682090f1cf742f8b2e88be35b3c6e2"} Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.309005 4969 scope.go:117] "RemoveContainer" containerID="87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.309179 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c25vm" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.342815 4969 scope.go:117] "RemoveContainer" containerID="e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.359371 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c25vm"] Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.370364 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c25vm"] Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.390235 4969 scope.go:117] "RemoveContainer" containerID="bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.433945 4969 scope.go:117] "RemoveContainer" containerID="87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb" Oct 04 09:12:20 crc kubenswrapper[4969]: E1004 09:12:20.434336 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb\": container with ID starting with 87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb not found: ID does not exist" containerID="87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.434367 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb"} err="failed to get container status \"87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb\": rpc error: code = NotFound desc = could not find container \"87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb\": container with ID starting with 87e8cb304645728332ca3089c1e59c80537c5722508218db726a8adf16f3fafb not found: ID does not exist" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.434391 4969 scope.go:117] "RemoveContainer" containerID="e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3" Oct 04 09:12:20 crc kubenswrapper[4969]: E1004 09:12:20.434668 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3\": container with ID starting with e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3 not found: ID does not exist" containerID="e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.434696 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3"} err="failed to get container status \"e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3\": rpc error: code = NotFound desc = could not find container \"e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3\": container with ID starting with e2ffd33abe6a81ec0cad4b8bce49e81efde7f00a7fbd69636cbf797e90258dd3 not found: ID does not exist" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.434716 4969 scope.go:117] "RemoveContainer" containerID="bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695" Oct 04 09:12:20 crc kubenswrapper[4969]: E1004 09:12:20.434974 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695\": container with ID starting with bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695 not found: ID does not exist" containerID="bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695" Oct 04 09:12:20 crc kubenswrapper[4969]: I1004 09:12:20.435002 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695"} err="failed to get container status \"bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695\": rpc error: code = NotFound desc = could not find container \"bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695\": container with ID starting with bb76ee77b9f37c3ef1d0b4bf7d2025a2f84366c0b9f1237d1403e0b7a72da695 not found: ID does not exist" Oct 04 09:12:21 crc kubenswrapper[4969]: I1004 09:12:21.071149 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" path="/var/lib/kubelet/pods/5e79a99d-a604-422f-9711-d8df00798f3c/volumes" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.653063 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rwb7v"] Oct 04 09:12:38 crc kubenswrapper[4969]: E1004 09:12:38.654406 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" containerName="registry-server" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.654455 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" containerName="registry-server" Oct 04 09:12:38 crc kubenswrapper[4969]: E1004 09:12:38.654496 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" containerName="extract-utilities" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.654513 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" containerName="extract-utilities" Oct 04 09:12:38 crc kubenswrapper[4969]: E1004 09:12:38.654557 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" containerName="extract-content" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.654571 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" containerName="extract-content" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.654920 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e79a99d-a604-422f-9711-d8df00798f3c" containerName="registry-server" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.657185 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.677214 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rwb7v"] Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.788217 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2hpf\" (UniqueName: \"kubernetes.io/projected/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-kube-api-access-l2hpf\") pod \"redhat-marketplace-rwb7v\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.788332 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-utilities\") pod \"redhat-marketplace-rwb7v\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.788501 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-catalog-content\") pod \"redhat-marketplace-rwb7v\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.890564 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2hpf\" (UniqueName: \"kubernetes.io/projected/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-kube-api-access-l2hpf\") pod \"redhat-marketplace-rwb7v\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.890909 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-utilities\") pod \"redhat-marketplace-rwb7v\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.891055 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-catalog-content\") pod \"redhat-marketplace-rwb7v\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.891380 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-utilities\") pod \"redhat-marketplace-rwb7v\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.891582 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-catalog-content\") pod \"redhat-marketplace-rwb7v\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:38 crc kubenswrapper[4969]: I1004 09:12:38.912042 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2hpf\" (UniqueName: \"kubernetes.io/projected/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-kube-api-access-l2hpf\") pod \"redhat-marketplace-rwb7v\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:39 crc kubenswrapper[4969]: I1004 09:12:39.027239 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:39 crc kubenswrapper[4969]: I1004 09:12:39.526057 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rwb7v"] Oct 04 09:12:40 crc kubenswrapper[4969]: I1004 09:12:40.560149 4969 generic.go:334] "Generic (PLEG): container finished" podID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerID="92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5" exitCode=0 Oct 04 09:12:40 crc kubenswrapper[4969]: I1004 09:12:40.560250 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rwb7v" event={"ID":"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe","Type":"ContainerDied","Data":"92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5"} Oct 04 09:12:40 crc kubenswrapper[4969]: I1004 09:12:40.563273 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rwb7v" event={"ID":"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe","Type":"ContainerStarted","Data":"cd369cebcde92225db39b0c14df7033f2bf14e77af7d3b37adf282ed5923e044"} Oct 04 09:12:42 crc kubenswrapper[4969]: I1004 09:12:42.603145 4969 generic.go:334] "Generic (PLEG): container finished" podID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerID="2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee" exitCode=0 Oct 04 09:12:42 crc kubenswrapper[4969]: I1004 09:12:42.603225 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rwb7v" event={"ID":"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe","Type":"ContainerDied","Data":"2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee"} Oct 04 09:12:43 crc kubenswrapper[4969]: I1004 09:12:43.615688 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rwb7v" event={"ID":"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe","Type":"ContainerStarted","Data":"fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27"} Oct 04 09:12:43 crc kubenswrapper[4969]: I1004 09:12:43.634955 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rwb7v" podStartSLOduration=3.129871655 podStartE2EDuration="5.634929852s" podCreationTimestamp="2025-10-04 09:12:38 +0000 UTC" firstStartedPulling="2025-10-04 09:12:40.562838139 +0000 UTC m=+3388.317106963" lastFinishedPulling="2025-10-04 09:12:43.067896326 +0000 UTC m=+3390.822165160" observedRunningTime="2025-10-04 09:12:43.632851691 +0000 UTC m=+3391.387120515" watchObservedRunningTime="2025-10-04 09:12:43.634929852 +0000 UTC m=+3391.389198686" Oct 04 09:12:49 crc kubenswrapper[4969]: I1004 09:12:49.028022 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:49 crc kubenswrapper[4969]: I1004 09:12:49.028766 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:49 crc kubenswrapper[4969]: I1004 09:12:49.128773 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:49 crc kubenswrapper[4969]: I1004 09:12:49.746967 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:49 crc kubenswrapper[4969]: I1004 09:12:49.808362 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rwb7v"] Oct 04 09:12:51 crc kubenswrapper[4969]: I1004 09:12:51.708857 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rwb7v" podUID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerName="registry-server" containerID="cri-o://fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27" gracePeriod=2 Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.304153 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.486280 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2hpf\" (UniqueName: \"kubernetes.io/projected/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-kube-api-access-l2hpf\") pod \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.486362 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-utilities\") pod \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.486384 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-catalog-content\") pod \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\" (UID: \"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe\") " Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.488070 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-utilities" (OuterVolumeSpecName: "utilities") pod "f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" (UID: "f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.497811 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-kube-api-access-l2hpf" (OuterVolumeSpecName: "kube-api-access-l2hpf") pod "f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" (UID: "f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe"). InnerVolumeSpecName "kube-api-access-l2hpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.504087 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" (UID: "f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.589275 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2hpf\" (UniqueName: \"kubernetes.io/projected/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-kube-api-access-l2hpf\") on node \"crc\" DevicePath \"\"" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.589318 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.589330 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.724822 4969 generic.go:334] "Generic (PLEG): container finished" podID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerID="fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27" exitCode=0 Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.724869 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rwb7v" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.724887 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rwb7v" event={"ID":"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe","Type":"ContainerDied","Data":"fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27"} Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.725896 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rwb7v" event={"ID":"f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe","Type":"ContainerDied","Data":"cd369cebcde92225db39b0c14df7033f2bf14e77af7d3b37adf282ed5923e044"} Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.725949 4969 scope.go:117] "RemoveContainer" containerID="fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.763039 4969 scope.go:117] "RemoveContainer" containerID="2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.788607 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rwb7v"] Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.822250 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rwb7v"] Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.845235 4969 scope.go:117] "RemoveContainer" containerID="92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.850956 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d77bw"] Oct 04 09:12:52 crc kubenswrapper[4969]: E1004 09:12:52.851405 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerName="registry-server" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.851438 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerName="registry-server" Oct 04 09:12:52 crc kubenswrapper[4969]: E1004 09:12:52.851461 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerName="extract-content" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.851470 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerName="extract-content" Oct 04 09:12:52 crc kubenswrapper[4969]: E1004 09:12:52.851493 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerName="extract-utilities" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.851501 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerName="extract-utilities" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.851796 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" containerName="registry-server" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.853392 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.866148 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d77bw"] Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.880445 4969 scope.go:117] "RemoveContainer" containerID="fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27" Oct 04 09:12:52 crc kubenswrapper[4969]: E1004 09:12:52.881988 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27\": container with ID starting with fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27 not found: ID does not exist" containerID="fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.882027 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27"} err="failed to get container status \"fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27\": rpc error: code = NotFound desc = could not find container \"fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27\": container with ID starting with fefe7e3329ea9a5385bc13eec002c324f8e1bc74198db27474cc89acd6e52f27 not found: ID does not exist" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.882059 4969 scope.go:117] "RemoveContainer" containerID="2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee" Oct 04 09:12:52 crc kubenswrapper[4969]: E1004 09:12:52.882569 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee\": container with ID starting with 2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee not found: ID does not exist" containerID="2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.882805 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee"} err="failed to get container status \"2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee\": rpc error: code = NotFound desc = could not find container \"2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee\": container with ID starting with 2d6bf7f0ba73da6b40605f61a9707965ff5fd394b545a1d2241e9a43a77cc6ee not found: ID does not exist" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.882841 4969 scope.go:117] "RemoveContainer" containerID="92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5" Oct 04 09:12:52 crc kubenswrapper[4969]: E1004 09:12:52.883105 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5\": container with ID starting with 92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5 not found: ID does not exist" containerID="92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.883132 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5"} err="failed to get container status \"92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5\": rpc error: code = NotFound desc = could not find container \"92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5\": container with ID starting with 92f963fc38bf105ebc000defe2f7e33f7756f133f7ab2c67d14f4204f81a42b5 not found: ID does not exist" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.997929 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmsp6\" (UniqueName: \"kubernetes.io/projected/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-kube-api-access-fmsp6\") pod \"redhat-operators-d77bw\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.998008 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-utilities\") pod \"redhat-operators-d77bw\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:52 crc kubenswrapper[4969]: I1004 09:12:52.998060 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-catalog-content\") pod \"redhat-operators-d77bw\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.068943 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe" path="/var/lib/kubelet/pods/f8f650cc-63b5-45a0-b4d1-68bf4f6c3dbe/volumes" Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.099809 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmsp6\" (UniqueName: \"kubernetes.io/projected/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-kube-api-access-fmsp6\") pod \"redhat-operators-d77bw\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.099866 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-utilities\") pod \"redhat-operators-d77bw\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.099906 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-catalog-content\") pod \"redhat-operators-d77bw\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.100692 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-utilities\") pod \"redhat-operators-d77bw\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.100763 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-catalog-content\") pod \"redhat-operators-d77bw\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.117181 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmsp6\" (UniqueName: \"kubernetes.io/projected/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-kube-api-access-fmsp6\") pod \"redhat-operators-d77bw\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.217596 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.676381 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d77bw"] Oct 04 09:12:53 crc kubenswrapper[4969]: W1004 09:12:53.682228 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcaeb304d_a31e_4aeb_9a1a_0f01f718eef1.slice/crio-a73079c36b49d90ee6f6ac99e4dc0850115cf316a2283b1c5c7a054f549f3bd2 WatchSource:0}: Error finding container a73079c36b49d90ee6f6ac99e4dc0850115cf316a2283b1c5c7a054f549f3bd2: Status 404 returned error can't find the container with id a73079c36b49d90ee6f6ac99e4dc0850115cf316a2283b1c5c7a054f549f3bd2 Oct 04 09:12:53 crc kubenswrapper[4969]: I1004 09:12:53.743863 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d77bw" event={"ID":"caeb304d-a31e-4aeb-9a1a-0f01f718eef1","Type":"ContainerStarted","Data":"a73079c36b49d90ee6f6ac99e4dc0850115cf316a2283b1c5c7a054f549f3bd2"} Oct 04 09:12:54 crc kubenswrapper[4969]: I1004 09:12:54.756567 4969 generic.go:334] "Generic (PLEG): container finished" podID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerID="1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4" exitCode=0 Oct 04 09:12:54 crc kubenswrapper[4969]: I1004 09:12:54.756692 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d77bw" event={"ID":"caeb304d-a31e-4aeb-9a1a-0f01f718eef1","Type":"ContainerDied","Data":"1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4"} Oct 04 09:12:56 crc kubenswrapper[4969]: I1004 09:12:56.785667 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d77bw" event={"ID":"caeb304d-a31e-4aeb-9a1a-0f01f718eef1","Type":"ContainerStarted","Data":"f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1"} Oct 04 09:13:01 crc kubenswrapper[4969]: I1004 09:13:01.848033 4969 generic.go:334] "Generic (PLEG): container finished" podID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerID="f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1" exitCode=0 Oct 04 09:13:01 crc kubenswrapper[4969]: I1004 09:13:01.848095 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d77bw" event={"ID":"caeb304d-a31e-4aeb-9a1a-0f01f718eef1","Type":"ContainerDied","Data":"f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1"} Oct 04 09:13:02 crc kubenswrapper[4969]: I1004 09:13:02.861962 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d77bw" event={"ID":"caeb304d-a31e-4aeb-9a1a-0f01f718eef1","Type":"ContainerStarted","Data":"e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924"} Oct 04 09:13:02 crc kubenswrapper[4969]: I1004 09:13:02.898599 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d77bw" podStartSLOduration=3.266710458 podStartE2EDuration="10.898566827s" podCreationTimestamp="2025-10-04 09:12:52 +0000 UTC" firstStartedPulling="2025-10-04 09:12:54.758654506 +0000 UTC m=+3402.512923330" lastFinishedPulling="2025-10-04 09:13:02.390510875 +0000 UTC m=+3410.144779699" observedRunningTime="2025-10-04 09:13:02.887178185 +0000 UTC m=+3410.641447059" watchObservedRunningTime="2025-10-04 09:13:02.898566827 +0000 UTC m=+3410.652835661" Oct 04 09:13:03 crc kubenswrapper[4969]: I1004 09:13:03.218516 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:13:03 crc kubenswrapper[4969]: I1004 09:13:03.218580 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:13:04 crc kubenswrapper[4969]: I1004 09:13:04.291027 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d77bw" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="registry-server" probeResult="failure" output=< Oct 04 09:13:04 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:13:04 crc kubenswrapper[4969]: > Oct 04 09:13:14 crc kubenswrapper[4969]: I1004 09:13:14.282684 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d77bw" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="registry-server" probeResult="failure" output=< Oct 04 09:13:14 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:13:14 crc kubenswrapper[4969]: > Oct 04 09:13:24 crc kubenswrapper[4969]: I1004 09:13:24.261533 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d77bw" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="registry-server" probeResult="failure" output=< Oct 04 09:13:24 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:13:24 crc kubenswrapper[4969]: > Oct 04 09:13:33 crc kubenswrapper[4969]: I1004 09:13:33.263266 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:13:33 crc kubenswrapper[4969]: I1004 09:13:33.310554 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:13:33 crc kubenswrapper[4969]: I1004 09:13:33.500475 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d77bw"] Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.244502 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d77bw" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="registry-server" containerID="cri-o://e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924" gracePeriod=2 Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.716917 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.820073 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-utilities\") pod \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.820119 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-catalog-content\") pod \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.820329 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmsp6\" (UniqueName: \"kubernetes.io/projected/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-kube-api-access-fmsp6\") pod \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\" (UID: \"caeb304d-a31e-4aeb-9a1a-0f01f718eef1\") " Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.821248 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-utilities" (OuterVolumeSpecName: "utilities") pod "caeb304d-a31e-4aeb-9a1a-0f01f718eef1" (UID: "caeb304d-a31e-4aeb-9a1a-0f01f718eef1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.826743 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-kube-api-access-fmsp6" (OuterVolumeSpecName: "kube-api-access-fmsp6") pod "caeb304d-a31e-4aeb-9a1a-0f01f718eef1" (UID: "caeb304d-a31e-4aeb-9a1a-0f01f718eef1"). InnerVolumeSpecName "kube-api-access-fmsp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.889878 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "caeb304d-a31e-4aeb-9a1a-0f01f718eef1" (UID: "caeb304d-a31e-4aeb-9a1a-0f01f718eef1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.924837 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.925183 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:13:35 crc kubenswrapper[4969]: I1004 09:13:35.925198 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmsp6\" (UniqueName: \"kubernetes.io/projected/caeb304d-a31e-4aeb-9a1a-0f01f718eef1-kube-api-access-fmsp6\") on node \"crc\" DevicePath \"\"" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.261660 4969 generic.go:334] "Generic (PLEG): container finished" podID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerID="e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924" exitCode=0 Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.261758 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d77bw" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.261742 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d77bw" event={"ID":"caeb304d-a31e-4aeb-9a1a-0f01f718eef1","Type":"ContainerDied","Data":"e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924"} Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.262032 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d77bw" event={"ID":"caeb304d-a31e-4aeb-9a1a-0f01f718eef1","Type":"ContainerDied","Data":"a73079c36b49d90ee6f6ac99e4dc0850115cf316a2283b1c5c7a054f549f3bd2"} Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.262087 4969 scope.go:117] "RemoveContainer" containerID="e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.309973 4969 scope.go:117] "RemoveContainer" containerID="f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.311778 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d77bw"] Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.321710 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d77bw"] Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.334824 4969 scope.go:117] "RemoveContainer" containerID="1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.389783 4969 scope.go:117] "RemoveContainer" containerID="e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924" Oct 04 09:13:36 crc kubenswrapper[4969]: E1004 09:13:36.390297 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924\": container with ID starting with e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924 not found: ID does not exist" containerID="e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.390354 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924"} err="failed to get container status \"e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924\": rpc error: code = NotFound desc = could not find container \"e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924\": container with ID starting with e633c5dd5eb4fe0aba49e9326fc7a7345a25464c3d73cfd6bfe18e598c269924 not found: ID does not exist" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.390386 4969 scope.go:117] "RemoveContainer" containerID="f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1" Oct 04 09:13:36 crc kubenswrapper[4969]: E1004 09:13:36.390783 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1\": container with ID starting with f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1 not found: ID does not exist" containerID="f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.390822 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1"} err="failed to get container status \"f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1\": rpc error: code = NotFound desc = could not find container \"f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1\": container with ID starting with f27f3d38250ab709491100eaa6f68c05d52da19b7636a13520ec37c1359d37f1 not found: ID does not exist" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.390843 4969 scope.go:117] "RemoveContainer" containerID="1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4" Oct 04 09:13:36 crc kubenswrapper[4969]: E1004 09:13:36.391127 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4\": container with ID starting with 1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4 not found: ID does not exist" containerID="1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4" Oct 04 09:13:36 crc kubenswrapper[4969]: I1004 09:13:36.391165 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4"} err="failed to get container status \"1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4\": rpc error: code = NotFound desc = could not find container \"1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4\": container with ID starting with 1c21ada08fda5e59b19102d7bfd0f08099e6fb1718fede39956d9fe6aad937c4 not found: ID does not exist" Oct 04 09:13:37 crc kubenswrapper[4969]: I1004 09:13:37.066374 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" path="/var/lib/kubelet/pods/caeb304d-a31e-4aeb-9a1a-0f01f718eef1/volumes" Oct 04 09:14:19 crc kubenswrapper[4969]: I1004 09:14:19.666938 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:14:19 crc kubenswrapper[4969]: I1004 09:14:19.667651 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:14:49 crc kubenswrapper[4969]: I1004 09:14:49.666220 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:14:49 crc kubenswrapper[4969]: I1004 09:14:49.666847 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.218464 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln"] Oct 04 09:15:00 crc kubenswrapper[4969]: E1004 09:15:00.219771 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="extract-utilities" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.219795 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="extract-utilities" Oct 04 09:15:00 crc kubenswrapper[4969]: E1004 09:15:00.219840 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="extract-content" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.219851 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="extract-content" Oct 04 09:15:00 crc kubenswrapper[4969]: E1004 09:15:00.219873 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="registry-server" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.219883 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="registry-server" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.220189 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="caeb304d-a31e-4aeb-9a1a-0f01f718eef1" containerName="registry-server" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.221135 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.223907 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.223916 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.232006 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln"] Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.312079 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-config-volume\") pod \"collect-profiles-29326155-hq9ln\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.312432 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjtm4\" (UniqueName: \"kubernetes.io/projected/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-kube-api-access-hjtm4\") pod \"collect-profiles-29326155-hq9ln\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.313089 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-secret-volume\") pod \"collect-profiles-29326155-hq9ln\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.415132 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjtm4\" (UniqueName: \"kubernetes.io/projected/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-kube-api-access-hjtm4\") pod \"collect-profiles-29326155-hq9ln\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.415615 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-secret-volume\") pod \"collect-profiles-29326155-hq9ln\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.415807 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-config-volume\") pod \"collect-profiles-29326155-hq9ln\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.417934 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-config-volume\") pod \"collect-profiles-29326155-hq9ln\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.441272 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-secret-volume\") pod \"collect-profiles-29326155-hq9ln\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.455722 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjtm4\" (UniqueName: \"kubernetes.io/projected/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-kube-api-access-hjtm4\") pod \"collect-profiles-29326155-hq9ln\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:00 crc kubenswrapper[4969]: I1004 09:15:00.546807 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:01 crc kubenswrapper[4969]: W1004 09:15:01.100282 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6243bf2a_5d33_41f1_a09f_bd0e441fbe4e.slice/crio-08265aae3a457bb82f786776b54730ff506e3cbabe5b571c264c975649731f6f WatchSource:0}: Error finding container 08265aae3a457bb82f786776b54730ff506e3cbabe5b571c264c975649731f6f: Status 404 returned error can't find the container with id 08265aae3a457bb82f786776b54730ff506e3cbabe5b571c264c975649731f6f Oct 04 09:15:01 crc kubenswrapper[4969]: I1004 09:15:01.100818 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln"] Oct 04 09:15:01 crc kubenswrapper[4969]: I1004 09:15:01.265810 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" event={"ID":"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e","Type":"ContainerStarted","Data":"08265aae3a457bb82f786776b54730ff506e3cbabe5b571c264c975649731f6f"} Oct 04 09:15:01 crc kubenswrapper[4969]: E1004 09:15:01.710110 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6243bf2a_5d33_41f1_a09f_bd0e441fbe4e.slice/crio-8e364dbc4824ba12e15260a6705964c53cb2061225c512fd1833f6d786805f10.scope\": RecentStats: unable to find data in memory cache]" Oct 04 09:15:02 crc kubenswrapper[4969]: I1004 09:15:02.277280 4969 generic.go:334] "Generic (PLEG): container finished" podID="6243bf2a-5d33-41f1-a09f-bd0e441fbe4e" containerID="8e364dbc4824ba12e15260a6705964c53cb2061225c512fd1833f6d786805f10" exitCode=0 Oct 04 09:15:02 crc kubenswrapper[4969]: I1004 09:15:02.277338 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" event={"ID":"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e","Type":"ContainerDied","Data":"8e364dbc4824ba12e15260a6705964c53cb2061225c512fd1833f6d786805f10"} Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.781590 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.886647 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjtm4\" (UniqueName: \"kubernetes.io/projected/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-kube-api-access-hjtm4\") pod \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.886699 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-secret-volume\") pod \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.887029 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-config-volume\") pod \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\" (UID: \"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e\") " Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.887980 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-config-volume" (OuterVolumeSpecName: "config-volume") pod "6243bf2a-5d33-41f1-a09f-bd0e441fbe4e" (UID: "6243bf2a-5d33-41f1-a09f-bd0e441fbe4e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.893449 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6243bf2a-5d33-41f1-a09f-bd0e441fbe4e" (UID: "6243bf2a-5d33-41f1-a09f-bd0e441fbe4e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.895076 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-kube-api-access-hjtm4" (OuterVolumeSpecName: "kube-api-access-hjtm4") pod "6243bf2a-5d33-41f1-a09f-bd0e441fbe4e" (UID: "6243bf2a-5d33-41f1-a09f-bd0e441fbe4e"). InnerVolumeSpecName "kube-api-access-hjtm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.990143 4969 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.990213 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjtm4\" (UniqueName: \"kubernetes.io/projected/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-kube-api-access-hjtm4\") on node \"crc\" DevicePath \"\"" Oct 04 09:15:03 crc kubenswrapper[4969]: I1004 09:15:03.990248 4969 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 09:15:04 crc kubenswrapper[4969]: I1004 09:15:04.320857 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" event={"ID":"6243bf2a-5d33-41f1-a09f-bd0e441fbe4e","Type":"ContainerDied","Data":"08265aae3a457bb82f786776b54730ff506e3cbabe5b571c264c975649731f6f"} Oct 04 09:15:04 crc kubenswrapper[4969]: I1004 09:15:04.320922 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08265aae3a457bb82f786776b54730ff506e3cbabe5b571c264c975649731f6f" Oct 04 09:15:04 crc kubenswrapper[4969]: I1004 09:15:04.321034 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln" Oct 04 09:15:04 crc kubenswrapper[4969]: I1004 09:15:04.885869 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs"] Oct 04 09:15:04 crc kubenswrapper[4969]: I1004 09:15:04.900891 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326110-txvxs"] Oct 04 09:15:09 crc kubenswrapper[4969]: I1004 09:15:09.315379 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c57b46a1-a938-4dc6-99c0-23c044e307da" path="/var/lib/kubelet/pods/c57b46a1-a938-4dc6-99c0-23c044e307da/volumes" Oct 04 09:15:09 crc kubenswrapper[4969]: E1004 09:15:09.333961 4969 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.28s" Oct 04 09:15:19 crc kubenswrapper[4969]: I1004 09:15:19.666370 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:15:19 crc kubenswrapper[4969]: I1004 09:15:19.667465 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:15:19 crc kubenswrapper[4969]: I1004 09:15:19.667565 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:15:19 crc kubenswrapper[4969]: I1004 09:15:19.669004 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6996838e394ee153292731315425104062d2c84f42bbc656b3b6c05ce870ab37"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:15:19 crc kubenswrapper[4969]: I1004 09:15:19.669137 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://6996838e394ee153292731315425104062d2c84f42bbc656b3b6c05ce870ab37" gracePeriod=600 Oct 04 09:15:20 crc kubenswrapper[4969]: I1004 09:15:20.553247 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="6996838e394ee153292731315425104062d2c84f42bbc656b3b6c05ce870ab37" exitCode=0 Oct 04 09:15:20 crc kubenswrapper[4969]: I1004 09:15:20.553363 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"6996838e394ee153292731315425104062d2c84f42bbc656b3b6c05ce870ab37"} Oct 04 09:15:20 crc kubenswrapper[4969]: I1004 09:15:20.553712 4969 scope.go:117] "RemoveContainer" containerID="47eb185349deaf6ea2007b3a446bfc7998f3e09e86fce3b6c52fa084a5fbefc2" Oct 04 09:15:21 crc kubenswrapper[4969]: I1004 09:15:21.574363 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29"} Oct 04 09:15:33 crc kubenswrapper[4969]: I1004 09:15:33.460811 4969 scope.go:117] "RemoveContainer" containerID="237c72f9d595bc3676d27b336cb77671d0283fe37474435a4c51d760fcf105c4" Oct 04 09:17:49 crc kubenswrapper[4969]: I1004 09:17:49.666704 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:17:49 crc kubenswrapper[4969]: I1004 09:17:49.667547 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:18:19 crc kubenswrapper[4969]: I1004 09:18:19.666492 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:18:19 crc kubenswrapper[4969]: I1004 09:18:19.666988 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:18:49 crc kubenswrapper[4969]: I1004 09:18:49.666371 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:18:49 crc kubenswrapper[4969]: I1004 09:18:49.667100 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:18:49 crc kubenswrapper[4969]: I1004 09:18:49.667167 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:18:49 crc kubenswrapper[4969]: I1004 09:18:49.668207 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:18:49 crc kubenswrapper[4969]: I1004 09:18:49.668300 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" gracePeriod=600 Oct 04 09:18:49 crc kubenswrapper[4969]: E1004 09:18:49.808758 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:18:50 crc kubenswrapper[4969]: I1004 09:18:50.150572 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" exitCode=0 Oct 04 09:18:50 crc kubenswrapper[4969]: I1004 09:18:50.150665 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29"} Oct 04 09:18:50 crc kubenswrapper[4969]: I1004 09:18:50.150851 4969 scope.go:117] "RemoveContainer" containerID="6996838e394ee153292731315425104062d2c84f42bbc656b3b6c05ce870ab37" Oct 04 09:18:50 crc kubenswrapper[4969]: I1004 09:18:50.151619 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:18:50 crc kubenswrapper[4969]: E1004 09:18:50.151948 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:19:03 crc kubenswrapper[4969]: I1004 09:19:03.064653 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:19:03 crc kubenswrapper[4969]: E1004 09:19:03.065757 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:19:18 crc kubenswrapper[4969]: I1004 09:19:18.055948 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:19:18 crc kubenswrapper[4969]: E1004 09:19:18.056781 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:19:30 crc kubenswrapper[4969]: I1004 09:19:30.055122 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:19:30 crc kubenswrapper[4969]: E1004 09:19:30.055950 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:19:41 crc kubenswrapper[4969]: I1004 09:19:41.057064 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:19:41 crc kubenswrapper[4969]: E1004 09:19:41.058452 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:19:52 crc kubenswrapper[4969]: I1004 09:19:52.055657 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:19:52 crc kubenswrapper[4969]: E1004 09:19:52.056740 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:20:06 crc kubenswrapper[4969]: I1004 09:20:06.055558 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:20:06 crc kubenswrapper[4969]: E1004 09:20:06.056775 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:20:21 crc kubenswrapper[4969]: I1004 09:20:21.055053 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:20:21 crc kubenswrapper[4969]: E1004 09:20:21.056079 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:20:32 crc kubenswrapper[4969]: I1004 09:20:32.056218 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:20:32 crc kubenswrapper[4969]: E1004 09:20:32.057242 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.155208 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pzn89"] Oct 04 09:20:41 crc kubenswrapper[4969]: E1004 09:20:41.156339 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6243bf2a-5d33-41f1-a09f-bd0e441fbe4e" containerName="collect-profiles" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.156356 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6243bf2a-5d33-41f1-a09f-bd0e441fbe4e" containerName="collect-profiles" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.156685 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="6243bf2a-5d33-41f1-a09f-bd0e441fbe4e" containerName="collect-profiles" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.158543 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.172074 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pzn89"] Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.241018 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-utilities\") pod \"certified-operators-pzn89\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.241079 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-catalog-content\") pod \"certified-operators-pzn89\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.241352 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2hpt\" (UniqueName: \"kubernetes.io/projected/e22e6e78-652f-4084-b541-2a7a1b6c636d-kube-api-access-n2hpt\") pod \"certified-operators-pzn89\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.343278 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-utilities\") pod \"certified-operators-pzn89\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.343315 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-catalog-content\") pod \"certified-operators-pzn89\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.345860 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2hpt\" (UniqueName: \"kubernetes.io/projected/e22e6e78-652f-4084-b541-2a7a1b6c636d-kube-api-access-n2hpt\") pod \"certified-operators-pzn89\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.346611 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-utilities\") pod \"certified-operators-pzn89\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.347159 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-catalog-content\") pod \"certified-operators-pzn89\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.387328 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2hpt\" (UniqueName: \"kubernetes.io/projected/e22e6e78-652f-4084-b541-2a7a1b6c636d-kube-api-access-n2hpt\") pod \"certified-operators-pzn89\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.481516 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:41 crc kubenswrapper[4969]: I1004 09:20:41.917693 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pzn89"] Oct 04 09:20:42 crc kubenswrapper[4969]: I1004 09:20:42.418600 4969 generic.go:334] "Generic (PLEG): container finished" podID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerID="337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08" exitCode=0 Oct 04 09:20:42 crc kubenswrapper[4969]: I1004 09:20:42.418749 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzn89" event={"ID":"e22e6e78-652f-4084-b541-2a7a1b6c636d","Type":"ContainerDied","Data":"337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08"} Oct 04 09:20:42 crc kubenswrapper[4969]: I1004 09:20:42.418967 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzn89" event={"ID":"e22e6e78-652f-4084-b541-2a7a1b6c636d","Type":"ContainerStarted","Data":"ff786805498a33d528bda10659c5c48f5b69ea5b67956e5428df3ed86d060d40"} Oct 04 09:20:42 crc kubenswrapper[4969]: I1004 09:20:42.421384 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 09:20:43 crc kubenswrapper[4969]: I1004 09:20:43.063334 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:20:43 crc kubenswrapper[4969]: E1004 09:20:43.064014 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:20:43 crc kubenswrapper[4969]: I1004 09:20:43.431757 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzn89" event={"ID":"e22e6e78-652f-4084-b541-2a7a1b6c636d","Type":"ContainerStarted","Data":"9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce"} Oct 04 09:20:46 crc kubenswrapper[4969]: I1004 09:20:46.468401 4969 generic.go:334] "Generic (PLEG): container finished" podID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerID="9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce" exitCode=0 Oct 04 09:20:46 crc kubenswrapper[4969]: I1004 09:20:46.468517 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzn89" event={"ID":"e22e6e78-652f-4084-b541-2a7a1b6c636d","Type":"ContainerDied","Data":"9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce"} Oct 04 09:20:48 crc kubenswrapper[4969]: I1004 09:20:48.505783 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzn89" event={"ID":"e22e6e78-652f-4084-b541-2a7a1b6c636d","Type":"ContainerStarted","Data":"12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5"} Oct 04 09:20:48 crc kubenswrapper[4969]: I1004 09:20:48.535659 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pzn89" podStartSLOduration=2.6086787 podStartE2EDuration="7.53563768s" podCreationTimestamp="2025-10-04 09:20:41 +0000 UTC" firstStartedPulling="2025-10-04 09:20:42.420994949 +0000 UTC m=+3870.175263803" lastFinishedPulling="2025-10-04 09:20:47.347953969 +0000 UTC m=+3875.102222783" observedRunningTime="2025-10-04 09:20:48.529306733 +0000 UTC m=+3876.283575547" watchObservedRunningTime="2025-10-04 09:20:48.53563768 +0000 UTC m=+3876.289906494" Oct 04 09:20:51 crc kubenswrapper[4969]: I1004 09:20:51.481999 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:51 crc kubenswrapper[4969]: I1004 09:20:51.484632 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:51 crc kubenswrapper[4969]: I1004 09:20:51.551747 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:20:54 crc kubenswrapper[4969]: I1004 09:20:54.056114 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:20:54 crc kubenswrapper[4969]: E1004 09:20:54.057183 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:21:01 crc kubenswrapper[4969]: I1004 09:21:01.556446 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:21:01 crc kubenswrapper[4969]: I1004 09:21:01.616139 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pzn89"] Oct 04 09:21:01 crc kubenswrapper[4969]: I1004 09:21:01.657993 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pzn89" podUID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerName="registry-server" containerID="cri-o://12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5" gracePeriod=2 Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.255095 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.304320 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-utilities\") pod \"e22e6e78-652f-4084-b541-2a7a1b6c636d\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.304713 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-catalog-content\") pod \"e22e6e78-652f-4084-b541-2a7a1b6c636d\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.304772 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2hpt\" (UniqueName: \"kubernetes.io/projected/e22e6e78-652f-4084-b541-2a7a1b6c636d-kube-api-access-n2hpt\") pod \"e22e6e78-652f-4084-b541-2a7a1b6c636d\" (UID: \"e22e6e78-652f-4084-b541-2a7a1b6c636d\") " Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.305548 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-utilities" (OuterVolumeSpecName: "utilities") pod "e22e6e78-652f-4084-b541-2a7a1b6c636d" (UID: "e22e6e78-652f-4084-b541-2a7a1b6c636d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.312032 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e22e6e78-652f-4084-b541-2a7a1b6c636d-kube-api-access-n2hpt" (OuterVolumeSpecName: "kube-api-access-n2hpt") pod "e22e6e78-652f-4084-b541-2a7a1b6c636d" (UID: "e22e6e78-652f-4084-b541-2a7a1b6c636d"). InnerVolumeSpecName "kube-api-access-n2hpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.374244 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e22e6e78-652f-4084-b541-2a7a1b6c636d" (UID: "e22e6e78-652f-4084-b541-2a7a1b6c636d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.406478 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.406508 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2hpt\" (UniqueName: \"kubernetes.io/projected/e22e6e78-652f-4084-b541-2a7a1b6c636d-kube-api-access-n2hpt\") on node \"crc\" DevicePath \"\"" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.406520 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e22e6e78-652f-4084-b541-2a7a1b6c636d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.676268 4969 generic.go:334] "Generic (PLEG): container finished" podID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerID="12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5" exitCode=0 Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.676328 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzn89" event={"ID":"e22e6e78-652f-4084-b541-2a7a1b6c636d","Type":"ContainerDied","Data":"12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5"} Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.676351 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzn89" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.676383 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzn89" event={"ID":"e22e6e78-652f-4084-b541-2a7a1b6c636d","Type":"ContainerDied","Data":"ff786805498a33d528bda10659c5c48f5b69ea5b67956e5428df3ed86d060d40"} Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.676507 4969 scope.go:117] "RemoveContainer" containerID="12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.701100 4969 scope.go:117] "RemoveContainer" containerID="9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.739966 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pzn89"] Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.745677 4969 scope.go:117] "RemoveContainer" containerID="337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.751376 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pzn89"] Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.811554 4969 scope.go:117] "RemoveContainer" containerID="12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5" Oct 04 09:21:02 crc kubenswrapper[4969]: E1004 09:21:02.812148 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5\": container with ID starting with 12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5 not found: ID does not exist" containerID="12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.812186 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5"} err="failed to get container status \"12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5\": rpc error: code = NotFound desc = could not find container \"12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5\": container with ID starting with 12c50d5f8077918e8047b10fa23f11adb50dda56b818596b2e11d7c0029593e5 not found: ID does not exist" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.812232 4969 scope.go:117] "RemoveContainer" containerID="9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce" Oct 04 09:21:02 crc kubenswrapper[4969]: E1004 09:21:02.812715 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce\": container with ID starting with 9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce not found: ID does not exist" containerID="9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.812758 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce"} err="failed to get container status \"9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce\": rpc error: code = NotFound desc = could not find container \"9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce\": container with ID starting with 9e5401f1ca1e0a0e78a16027ac84a0e4b4e1acfbdf7088947280b21c3f3489ce not found: ID does not exist" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.812790 4969 scope.go:117] "RemoveContainer" containerID="337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08" Oct 04 09:21:02 crc kubenswrapper[4969]: E1004 09:21:02.814840 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08\": container with ID starting with 337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08 not found: ID does not exist" containerID="337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08" Oct 04 09:21:02 crc kubenswrapper[4969]: I1004 09:21:02.814939 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08"} err="failed to get container status \"337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08\": rpc error: code = NotFound desc = could not find container \"337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08\": container with ID starting with 337178ed86622dd8986fc72f8b7fa29a0df9196c9d3c419bf4808a3209e6ae08 not found: ID does not exist" Oct 04 09:21:03 crc kubenswrapper[4969]: I1004 09:21:03.072765 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e22e6e78-652f-4084-b541-2a7a1b6c636d" path="/var/lib/kubelet/pods/e22e6e78-652f-4084-b541-2a7a1b6c636d/volumes" Oct 04 09:21:08 crc kubenswrapper[4969]: I1004 09:21:08.056786 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:21:08 crc kubenswrapper[4969]: E1004 09:21:08.059710 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:21:21 crc kubenswrapper[4969]: I1004 09:21:21.056991 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:21:21 crc kubenswrapper[4969]: E1004 09:21:21.058134 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:21:32 crc kubenswrapper[4969]: I1004 09:21:32.054699 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:21:32 crc kubenswrapper[4969]: E1004 09:21:32.055405 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:21:46 crc kubenswrapper[4969]: I1004 09:21:46.056331 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:21:46 crc kubenswrapper[4969]: E1004 09:21:46.057178 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:22:01 crc kubenswrapper[4969]: I1004 09:22:01.056154 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:22:01 crc kubenswrapper[4969]: E1004 09:22:01.057179 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:22:13 crc kubenswrapper[4969]: I1004 09:22:13.065836 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:22:13 crc kubenswrapper[4969]: E1004 09:22:13.067001 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.358908 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9c8ff"] Oct 04 09:22:20 crc kubenswrapper[4969]: E1004 09:22:20.373571 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerName="registry-server" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.373635 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerName="registry-server" Oct 04 09:22:20 crc kubenswrapper[4969]: E1004 09:22:20.373717 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerName="extract-utilities" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.373728 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerName="extract-utilities" Oct 04 09:22:20 crc kubenswrapper[4969]: E1004 09:22:20.373810 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerName="extract-content" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.373821 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerName="extract-content" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.374481 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22e6e78-652f-4084-b541-2a7a1b6c636d" containerName="registry-server" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.381977 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.396210 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9c8ff"] Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.456773 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-utilities\") pod \"community-operators-9c8ff\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.456917 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8htp\" (UniqueName: \"kubernetes.io/projected/eae99fd2-8614-43e2-9744-5b390b8ff47e-kube-api-access-j8htp\") pod \"community-operators-9c8ff\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.457090 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-catalog-content\") pod \"community-operators-9c8ff\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.558965 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-utilities\") pod \"community-operators-9c8ff\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.559066 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8htp\" (UniqueName: \"kubernetes.io/projected/eae99fd2-8614-43e2-9744-5b390b8ff47e-kube-api-access-j8htp\") pod \"community-operators-9c8ff\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.559164 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-catalog-content\") pod \"community-operators-9c8ff\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.559713 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-catalog-content\") pod \"community-operators-9c8ff\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.559807 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-utilities\") pod \"community-operators-9c8ff\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.585489 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8htp\" (UniqueName: \"kubernetes.io/projected/eae99fd2-8614-43e2-9744-5b390b8ff47e-kube-api-access-j8htp\") pod \"community-operators-9c8ff\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:20 crc kubenswrapper[4969]: I1004 09:22:20.712004 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:22 crc kubenswrapper[4969]: W1004 09:22:21.260406 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeae99fd2_8614_43e2_9744_5b390b8ff47e.slice/crio-54eedf4e5590afa30f29c02177c490c9d60cdf06e55afbbfd72a860ce2551f86 WatchSource:0}: Error finding container 54eedf4e5590afa30f29c02177c490c9d60cdf06e55afbbfd72a860ce2551f86: Status 404 returned error can't find the container with id 54eedf4e5590afa30f29c02177c490c9d60cdf06e55afbbfd72a860ce2551f86 Oct 04 09:22:22 crc kubenswrapper[4969]: I1004 09:22:21.262714 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9c8ff"] Oct 04 09:22:22 crc kubenswrapper[4969]: I1004 09:22:21.673186 4969 generic.go:334] "Generic (PLEG): container finished" podID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerID="cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f" exitCode=0 Oct 04 09:22:22 crc kubenswrapper[4969]: I1004 09:22:21.673241 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8ff" event={"ID":"eae99fd2-8614-43e2-9744-5b390b8ff47e","Type":"ContainerDied","Data":"cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f"} Oct 04 09:22:22 crc kubenswrapper[4969]: I1004 09:22:21.673279 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8ff" event={"ID":"eae99fd2-8614-43e2-9744-5b390b8ff47e","Type":"ContainerStarted","Data":"54eedf4e5590afa30f29c02177c490c9d60cdf06e55afbbfd72a860ce2551f86"} Oct 04 09:22:23 crc kubenswrapper[4969]: I1004 09:22:23.701342 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8ff" event={"ID":"eae99fd2-8614-43e2-9744-5b390b8ff47e","Type":"ContainerStarted","Data":"265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89"} Oct 04 09:22:26 crc kubenswrapper[4969]: I1004 09:22:26.738654 4969 generic.go:334] "Generic (PLEG): container finished" podID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerID="265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89" exitCode=0 Oct 04 09:22:26 crc kubenswrapper[4969]: I1004 09:22:26.738751 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8ff" event={"ID":"eae99fd2-8614-43e2-9744-5b390b8ff47e","Type":"ContainerDied","Data":"265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89"} Oct 04 09:22:27 crc kubenswrapper[4969]: I1004 09:22:27.056713 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:22:27 crc kubenswrapper[4969]: E1004 09:22:27.056978 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:22:27 crc kubenswrapper[4969]: I1004 09:22:27.752436 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8ff" event={"ID":"eae99fd2-8614-43e2-9744-5b390b8ff47e","Type":"ContainerStarted","Data":"5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8"} Oct 04 09:22:27 crc kubenswrapper[4969]: I1004 09:22:27.772957 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9c8ff" podStartSLOduration=2.224995002 podStartE2EDuration="7.772938192s" podCreationTimestamp="2025-10-04 09:22:20 +0000 UTC" firstStartedPulling="2025-10-04 09:22:21.675639519 +0000 UTC m=+3969.429908353" lastFinishedPulling="2025-10-04 09:22:27.223582689 +0000 UTC m=+3974.977851543" observedRunningTime="2025-10-04 09:22:27.770293556 +0000 UTC m=+3975.524562370" watchObservedRunningTime="2025-10-04 09:22:27.772938192 +0000 UTC m=+3975.527206996" Oct 04 09:22:30 crc kubenswrapper[4969]: I1004 09:22:30.713187 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:30 crc kubenswrapper[4969]: I1004 09:22:30.714579 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:30 crc kubenswrapper[4969]: I1004 09:22:30.785296 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:39 crc kubenswrapper[4969]: I1004 09:22:39.056161 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:22:39 crc kubenswrapper[4969]: E1004 09:22:39.057321 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:22:40 crc kubenswrapper[4969]: I1004 09:22:40.820745 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:40 crc kubenswrapper[4969]: I1004 09:22:40.903970 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9c8ff"] Oct 04 09:22:40 crc kubenswrapper[4969]: I1004 09:22:40.904217 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9c8ff" podUID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerName="registry-server" containerID="cri-o://5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8" gracePeriod=2 Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.445788 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.625418 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-catalog-content\") pod \"eae99fd2-8614-43e2-9744-5b390b8ff47e\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.625846 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8htp\" (UniqueName: \"kubernetes.io/projected/eae99fd2-8614-43e2-9744-5b390b8ff47e-kube-api-access-j8htp\") pod \"eae99fd2-8614-43e2-9744-5b390b8ff47e\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.625934 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-utilities\") pod \"eae99fd2-8614-43e2-9744-5b390b8ff47e\" (UID: \"eae99fd2-8614-43e2-9744-5b390b8ff47e\") " Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.626606 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-utilities" (OuterVolumeSpecName: "utilities") pod "eae99fd2-8614-43e2-9744-5b390b8ff47e" (UID: "eae99fd2-8614-43e2-9744-5b390b8ff47e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.630630 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae99fd2-8614-43e2-9744-5b390b8ff47e-kube-api-access-j8htp" (OuterVolumeSpecName: "kube-api-access-j8htp") pod "eae99fd2-8614-43e2-9744-5b390b8ff47e" (UID: "eae99fd2-8614-43e2-9744-5b390b8ff47e"). InnerVolumeSpecName "kube-api-access-j8htp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.670340 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eae99fd2-8614-43e2-9744-5b390b8ff47e" (UID: "eae99fd2-8614-43e2-9744-5b390b8ff47e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.728065 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8htp\" (UniqueName: \"kubernetes.io/projected/eae99fd2-8614-43e2-9744-5b390b8ff47e-kube-api-access-j8htp\") on node \"crc\" DevicePath \"\"" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.728097 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.728107 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae99fd2-8614-43e2-9744-5b390b8ff47e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.913033 4969 generic.go:334] "Generic (PLEG): container finished" podID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerID="5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8" exitCode=0 Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.913079 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8ff" event={"ID":"eae99fd2-8614-43e2-9744-5b390b8ff47e","Type":"ContainerDied","Data":"5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8"} Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.913094 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9c8ff" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.913106 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8ff" event={"ID":"eae99fd2-8614-43e2-9744-5b390b8ff47e","Type":"ContainerDied","Data":"54eedf4e5590afa30f29c02177c490c9d60cdf06e55afbbfd72a860ce2551f86"} Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.913123 4969 scope.go:117] "RemoveContainer" containerID="5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.934943 4969 scope.go:117] "RemoveContainer" containerID="265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89" Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.956227 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9c8ff"] Oct 04 09:22:41 crc kubenswrapper[4969]: I1004 09:22:41.974154 4969 scope.go:117] "RemoveContainer" containerID="cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f" Oct 04 09:22:42 crc kubenswrapper[4969]: I1004 09:22:42.005603 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9c8ff"] Oct 04 09:22:42 crc kubenswrapper[4969]: I1004 09:22:42.017600 4969 scope.go:117] "RemoveContainer" containerID="5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8" Oct 04 09:22:42 crc kubenswrapper[4969]: E1004 09:22:42.018221 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8\": container with ID starting with 5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8 not found: ID does not exist" containerID="5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8" Oct 04 09:22:42 crc kubenswrapper[4969]: I1004 09:22:42.018277 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8"} err="failed to get container status \"5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8\": rpc error: code = NotFound desc = could not find container \"5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8\": container with ID starting with 5cab42b5055801bf8193cebe3f4409885cdd84ef7d9f6141123d45f7b5205ed8 not found: ID does not exist" Oct 04 09:22:42 crc kubenswrapper[4969]: I1004 09:22:42.018315 4969 scope.go:117] "RemoveContainer" containerID="265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89" Oct 04 09:22:42 crc kubenswrapper[4969]: E1004 09:22:42.018775 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89\": container with ID starting with 265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89 not found: ID does not exist" containerID="265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89" Oct 04 09:22:42 crc kubenswrapper[4969]: I1004 09:22:42.018904 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89"} err="failed to get container status \"265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89\": rpc error: code = NotFound desc = could not find container \"265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89\": container with ID starting with 265bea9839ffb3bafb75bb02232143dea8f511e30bd3ec43c2100d25cef29b89 not found: ID does not exist" Oct 04 09:22:42 crc kubenswrapper[4969]: I1004 09:22:42.019010 4969 scope.go:117] "RemoveContainer" containerID="cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f" Oct 04 09:22:42 crc kubenswrapper[4969]: E1004 09:22:42.019398 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f\": container with ID starting with cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f not found: ID does not exist" containerID="cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f" Oct 04 09:22:42 crc kubenswrapper[4969]: I1004 09:22:42.019501 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f"} err="failed to get container status \"cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f\": rpc error: code = NotFound desc = could not find container \"cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f\": container with ID starting with cfaf56fca2c2518227a2563e814616199f6f9736c31762199fd77adffd32747f not found: ID does not exist" Oct 04 09:22:43 crc kubenswrapper[4969]: I1004 09:22:43.067526 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae99fd2-8614-43e2-9744-5b390b8ff47e" path="/var/lib/kubelet/pods/eae99fd2-8614-43e2-9744-5b390b8ff47e/volumes" Oct 04 09:22:54 crc kubenswrapper[4969]: I1004 09:22:54.055975 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:22:54 crc kubenswrapper[4969]: E1004 09:22:54.057075 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:23:05 crc kubenswrapper[4969]: I1004 09:23:05.055175 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:23:05 crc kubenswrapper[4969]: E1004 09:23:05.056118 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:23:17 crc kubenswrapper[4969]: I1004 09:23:17.055895 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:23:17 crc kubenswrapper[4969]: E1004 09:23:17.056658 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.250976 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wvx8r"] Oct 04 09:23:25 crc kubenswrapper[4969]: E1004 09:23:25.252260 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerName="extract-utilities" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.252286 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerName="extract-utilities" Oct 04 09:23:25 crc kubenswrapper[4969]: E1004 09:23:25.252314 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerName="registry-server" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.252326 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerName="registry-server" Oct 04 09:23:25 crc kubenswrapper[4969]: E1004 09:23:25.252372 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerName="extract-content" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.252384 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerName="extract-content" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.252767 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae99fd2-8614-43e2-9744-5b390b8ff47e" containerName="registry-server" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.255358 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.266323 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wvx8r"] Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.364411 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-catalog-content\") pod \"redhat-operators-wvx8r\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.364482 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-utilities\") pod \"redhat-operators-wvx8r\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.364517 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x846d\" (UniqueName: \"kubernetes.io/projected/be393ed6-50ce-49cf-bdd7-6add0751295d-kube-api-access-x846d\") pod \"redhat-operators-wvx8r\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.466033 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-catalog-content\") pod \"redhat-operators-wvx8r\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.466090 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-utilities\") pod \"redhat-operators-wvx8r\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.466154 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x846d\" (UniqueName: \"kubernetes.io/projected/be393ed6-50ce-49cf-bdd7-6add0751295d-kube-api-access-x846d\") pod \"redhat-operators-wvx8r\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.466604 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-catalog-content\") pod \"redhat-operators-wvx8r\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.466786 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-utilities\") pod \"redhat-operators-wvx8r\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.497176 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x846d\" (UniqueName: \"kubernetes.io/projected/be393ed6-50ce-49cf-bdd7-6add0751295d-kube-api-access-x846d\") pod \"redhat-operators-wvx8r\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:25 crc kubenswrapper[4969]: I1004 09:23:25.583900 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:26 crc kubenswrapper[4969]: I1004 09:23:26.042519 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wvx8r"] Oct 04 09:23:26 crc kubenswrapper[4969]: I1004 09:23:26.469044 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvx8r" event={"ID":"be393ed6-50ce-49cf-bdd7-6add0751295d","Type":"ContainerStarted","Data":"ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc"} Oct 04 09:23:26 crc kubenswrapper[4969]: I1004 09:23:26.469440 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvx8r" event={"ID":"be393ed6-50ce-49cf-bdd7-6add0751295d","Type":"ContainerStarted","Data":"d0d8d790af24bda84539918ae17db63430b00c22b9c3eddb0f1e2c1c9e6f8e2f"} Oct 04 09:23:27 crc kubenswrapper[4969]: I1004 09:23:27.488528 4969 generic.go:334] "Generic (PLEG): container finished" podID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerID="ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc" exitCode=0 Oct 04 09:23:27 crc kubenswrapper[4969]: I1004 09:23:27.488683 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvx8r" event={"ID":"be393ed6-50ce-49cf-bdd7-6add0751295d","Type":"ContainerDied","Data":"ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc"} Oct 04 09:23:29 crc kubenswrapper[4969]: I1004 09:23:29.513133 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvx8r" event={"ID":"be393ed6-50ce-49cf-bdd7-6add0751295d","Type":"ContainerStarted","Data":"9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617"} Oct 04 09:23:30 crc kubenswrapper[4969]: I1004 09:23:30.058011 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:23:30 crc kubenswrapper[4969]: E1004 09:23:30.062677 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:23:32 crc kubenswrapper[4969]: I1004 09:23:32.556760 4969 generic.go:334] "Generic (PLEG): container finished" podID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerID="9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617" exitCode=0 Oct 04 09:23:32 crc kubenswrapper[4969]: I1004 09:23:32.556872 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvx8r" event={"ID":"be393ed6-50ce-49cf-bdd7-6add0751295d","Type":"ContainerDied","Data":"9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617"} Oct 04 09:23:33 crc kubenswrapper[4969]: I1004 09:23:33.573923 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvx8r" event={"ID":"be393ed6-50ce-49cf-bdd7-6add0751295d","Type":"ContainerStarted","Data":"ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae"} Oct 04 09:23:33 crc kubenswrapper[4969]: I1004 09:23:33.604185 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wvx8r" podStartSLOduration=2.975413912 podStartE2EDuration="8.604153829s" podCreationTimestamp="2025-10-04 09:23:25 +0000 UTC" firstStartedPulling="2025-10-04 09:23:27.492179934 +0000 UTC m=+4035.246448778" lastFinishedPulling="2025-10-04 09:23:33.120919861 +0000 UTC m=+4040.875188695" observedRunningTime="2025-10-04 09:23:33.602197251 +0000 UTC m=+4041.356466075" watchObservedRunningTime="2025-10-04 09:23:33.604153829 +0000 UTC m=+4041.358422683" Oct 04 09:23:35 crc kubenswrapper[4969]: I1004 09:23:35.584304 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:35 crc kubenswrapper[4969]: I1004 09:23:35.584698 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:36 crc kubenswrapper[4969]: I1004 09:23:36.653665 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wvx8r" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerName="registry-server" probeResult="failure" output=< Oct 04 09:23:36 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:23:36 crc kubenswrapper[4969]: > Oct 04 09:23:41 crc kubenswrapper[4969]: I1004 09:23:41.055360 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:23:41 crc kubenswrapper[4969]: E1004 09:23:41.056106 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:23:45 crc kubenswrapper[4969]: I1004 09:23:45.634507 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:45 crc kubenswrapper[4969]: I1004 09:23:45.691180 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:45 crc kubenswrapper[4969]: I1004 09:23:45.876008 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wvx8r"] Oct 04 09:23:46 crc kubenswrapper[4969]: I1004 09:23:46.736395 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wvx8r" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerName="registry-server" containerID="cri-o://ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae" gracePeriod=2 Oct 04 09:23:46 crc kubenswrapper[4969]: E1004 09:23:46.967025 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe393ed6_50ce_49cf_bdd7_6add0751295d.slice/crio-conmon-ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae.scope\": RecentStats: unable to find data in memory cache]" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.364326 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.460535 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-utilities\") pod \"be393ed6-50ce-49cf-bdd7-6add0751295d\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.460717 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-catalog-content\") pod \"be393ed6-50ce-49cf-bdd7-6add0751295d\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.461070 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x846d\" (UniqueName: \"kubernetes.io/projected/be393ed6-50ce-49cf-bdd7-6add0751295d-kube-api-access-x846d\") pod \"be393ed6-50ce-49cf-bdd7-6add0751295d\" (UID: \"be393ed6-50ce-49cf-bdd7-6add0751295d\") " Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.464286 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-utilities" (OuterVolumeSpecName: "utilities") pod "be393ed6-50ce-49cf-bdd7-6add0751295d" (UID: "be393ed6-50ce-49cf-bdd7-6add0751295d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.471119 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be393ed6-50ce-49cf-bdd7-6add0751295d-kube-api-access-x846d" (OuterVolumeSpecName: "kube-api-access-x846d") pod "be393ed6-50ce-49cf-bdd7-6add0751295d" (UID: "be393ed6-50ce-49cf-bdd7-6add0751295d"). InnerVolumeSpecName "kube-api-access-x846d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.549251 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be393ed6-50ce-49cf-bdd7-6add0751295d" (UID: "be393ed6-50ce-49cf-bdd7-6add0751295d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.564177 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x846d\" (UniqueName: \"kubernetes.io/projected/be393ed6-50ce-49cf-bdd7-6add0751295d-kube-api-access-x846d\") on node \"crc\" DevicePath \"\"" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.564201 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.564212 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be393ed6-50ce-49cf-bdd7-6add0751295d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.755818 4969 generic.go:334] "Generic (PLEG): container finished" podID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerID="ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae" exitCode=0 Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.755842 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wvx8r" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.755864 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvx8r" event={"ID":"be393ed6-50ce-49cf-bdd7-6add0751295d","Type":"ContainerDied","Data":"ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae"} Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.756050 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvx8r" event={"ID":"be393ed6-50ce-49cf-bdd7-6add0751295d","Type":"ContainerDied","Data":"d0d8d790af24bda84539918ae17db63430b00c22b9c3eddb0f1e2c1c9e6f8e2f"} Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.756069 4969 scope.go:117] "RemoveContainer" containerID="ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.792095 4969 scope.go:117] "RemoveContainer" containerID="9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.818871 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wvx8r"] Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.833151 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wvx8r"] Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.849369 4969 scope.go:117] "RemoveContainer" containerID="ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.884355 4969 scope.go:117] "RemoveContainer" containerID="ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae" Oct 04 09:23:47 crc kubenswrapper[4969]: E1004 09:23:47.885001 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae\": container with ID starting with ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae not found: ID does not exist" containerID="ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.885058 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae"} err="failed to get container status \"ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae\": rpc error: code = NotFound desc = could not find container \"ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae\": container with ID starting with ba99445bad78caf5b1f6c96a0fc63a0c51e286bec0ed278d620080e41dde02ae not found: ID does not exist" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.885094 4969 scope.go:117] "RemoveContainer" containerID="9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617" Oct 04 09:23:47 crc kubenswrapper[4969]: E1004 09:23:47.885875 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617\": container with ID starting with 9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617 not found: ID does not exist" containerID="9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.885909 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617"} err="failed to get container status \"9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617\": rpc error: code = NotFound desc = could not find container \"9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617\": container with ID starting with 9a9c52126dae7cae9ca609f40e6f1c4c3508b9e8d2fa48fa089232c0f14ee617 not found: ID does not exist" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.885992 4969 scope.go:117] "RemoveContainer" containerID="ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc" Oct 04 09:23:47 crc kubenswrapper[4969]: E1004 09:23:47.886647 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc\": container with ID starting with ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc not found: ID does not exist" containerID="ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc" Oct 04 09:23:47 crc kubenswrapper[4969]: I1004 09:23:47.886699 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc"} err="failed to get container status \"ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc\": rpc error: code = NotFound desc = could not find container \"ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc\": container with ID starting with ad3ac330265c19f55484a53f94cfbb941268dd450dda10a1c351cb26aba281fc not found: ID does not exist" Oct 04 09:23:49 crc kubenswrapper[4969]: I1004 09:23:49.073835 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" path="/var/lib/kubelet/pods/be393ed6-50ce-49cf-bdd7-6add0751295d/volumes" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.287180 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xjv54"] Oct 04 09:23:50 crc kubenswrapper[4969]: E1004 09:23:50.288047 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerName="extract-utilities" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.288070 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerName="extract-utilities" Oct 04 09:23:50 crc kubenswrapper[4969]: E1004 09:23:50.288119 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerName="extract-content" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.288130 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerName="extract-content" Oct 04 09:23:50 crc kubenswrapper[4969]: E1004 09:23:50.288159 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerName="registry-server" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.288169 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerName="registry-server" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.288577 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="be393ed6-50ce-49cf-bdd7-6add0751295d" containerName="registry-server" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.290657 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.313300 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjv54"] Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.324847 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-utilities\") pod \"redhat-marketplace-xjv54\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.324959 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84dkd\" (UniqueName: \"kubernetes.io/projected/733d7f35-d085-4201-8f2d-d1452f410ad0-kube-api-access-84dkd\") pod \"redhat-marketplace-xjv54\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.325047 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-catalog-content\") pod \"redhat-marketplace-xjv54\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.427718 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84dkd\" (UniqueName: \"kubernetes.io/projected/733d7f35-d085-4201-8f2d-d1452f410ad0-kube-api-access-84dkd\") pod \"redhat-marketplace-xjv54\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.427838 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-catalog-content\") pod \"redhat-marketplace-xjv54\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.427981 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-utilities\") pod \"redhat-marketplace-xjv54\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.428445 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-catalog-content\") pod \"redhat-marketplace-xjv54\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.428535 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-utilities\") pod \"redhat-marketplace-xjv54\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.449225 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84dkd\" (UniqueName: \"kubernetes.io/projected/733d7f35-d085-4201-8f2d-d1452f410ad0-kube-api-access-84dkd\") pod \"redhat-marketplace-xjv54\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:50 crc kubenswrapper[4969]: I1004 09:23:50.618206 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:23:51 crc kubenswrapper[4969]: I1004 09:23:51.083282 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjv54"] Oct 04 09:23:51 crc kubenswrapper[4969]: I1004 09:23:51.808359 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjv54" event={"ID":"733d7f35-d085-4201-8f2d-d1452f410ad0","Type":"ContainerStarted","Data":"84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d"} Oct 04 09:23:51 crc kubenswrapper[4969]: I1004 09:23:51.808658 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjv54" event={"ID":"733d7f35-d085-4201-8f2d-d1452f410ad0","Type":"ContainerStarted","Data":"258b7972938ea6674f695896402498771a95465c3b9e4d557b29d6ebf59bed28"} Oct 04 09:23:52 crc kubenswrapper[4969]: I1004 09:23:52.824934 4969 generic.go:334] "Generic (PLEG): container finished" podID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerID="84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d" exitCode=0 Oct 04 09:23:52 crc kubenswrapper[4969]: I1004 09:23:52.825011 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjv54" event={"ID":"733d7f35-d085-4201-8f2d-d1452f410ad0","Type":"ContainerDied","Data":"84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d"} Oct 04 09:23:54 crc kubenswrapper[4969]: I1004 09:23:54.846632 4969 generic.go:334] "Generic (PLEG): container finished" podID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerID="c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386" exitCode=0 Oct 04 09:23:54 crc kubenswrapper[4969]: I1004 09:23:54.848193 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjv54" event={"ID":"733d7f35-d085-4201-8f2d-d1452f410ad0","Type":"ContainerDied","Data":"c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386"} Oct 04 09:23:55 crc kubenswrapper[4969]: I1004 09:23:55.055609 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:23:55 crc kubenswrapper[4969]: I1004 09:23:55.874785 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"baaf5b20707ce5ede797eb8602534733af5d13077a0e17e8fdf8e0c669c98383"} Oct 04 09:23:56 crc kubenswrapper[4969]: I1004 09:23:56.889486 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjv54" event={"ID":"733d7f35-d085-4201-8f2d-d1452f410ad0","Type":"ContainerStarted","Data":"e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb"} Oct 04 09:23:56 crc kubenswrapper[4969]: I1004 09:23:56.911915 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xjv54" podStartSLOduration=3.8976743259999997 podStartE2EDuration="6.911898432s" podCreationTimestamp="2025-10-04 09:23:50 +0000 UTC" firstStartedPulling="2025-10-04 09:23:52.82801861 +0000 UTC m=+4060.582287464" lastFinishedPulling="2025-10-04 09:23:55.842242756 +0000 UTC m=+4063.596511570" observedRunningTime="2025-10-04 09:23:56.907434832 +0000 UTC m=+4064.661703646" watchObservedRunningTime="2025-10-04 09:23:56.911898432 +0000 UTC m=+4064.666167236" Oct 04 09:24:00 crc kubenswrapper[4969]: I1004 09:24:00.618802 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:24:00 crc kubenswrapper[4969]: I1004 09:24:00.619668 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:24:00 crc kubenswrapper[4969]: I1004 09:24:00.684784 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:24:10 crc kubenswrapper[4969]: I1004 09:24:10.691144 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:24:10 crc kubenswrapper[4969]: I1004 09:24:10.751854 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjv54"] Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.062506 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xjv54" podUID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerName="registry-server" containerID="cri-o://e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb" gracePeriod=2 Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.602910 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.735223 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84dkd\" (UniqueName: \"kubernetes.io/projected/733d7f35-d085-4201-8f2d-d1452f410ad0-kube-api-access-84dkd\") pod \"733d7f35-d085-4201-8f2d-d1452f410ad0\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.735564 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-utilities\") pod \"733d7f35-d085-4201-8f2d-d1452f410ad0\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.735590 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-catalog-content\") pod \"733d7f35-d085-4201-8f2d-d1452f410ad0\" (UID: \"733d7f35-d085-4201-8f2d-d1452f410ad0\") " Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.736398 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-utilities" (OuterVolumeSpecName: "utilities") pod "733d7f35-d085-4201-8f2d-d1452f410ad0" (UID: "733d7f35-d085-4201-8f2d-d1452f410ad0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.746556 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/733d7f35-d085-4201-8f2d-d1452f410ad0-kube-api-access-84dkd" (OuterVolumeSpecName: "kube-api-access-84dkd") pod "733d7f35-d085-4201-8f2d-d1452f410ad0" (UID: "733d7f35-d085-4201-8f2d-d1452f410ad0"). InnerVolumeSpecName "kube-api-access-84dkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.748573 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "733d7f35-d085-4201-8f2d-d1452f410ad0" (UID: "733d7f35-d085-4201-8f2d-d1452f410ad0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.838893 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.838927 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733d7f35-d085-4201-8f2d-d1452f410ad0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:24:11 crc kubenswrapper[4969]: I1004 09:24:11.838942 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84dkd\" (UniqueName: \"kubernetes.io/projected/733d7f35-d085-4201-8f2d-d1452f410ad0-kube-api-access-84dkd\") on node \"crc\" DevicePath \"\"" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.079582 4969 generic.go:334] "Generic (PLEG): container finished" podID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerID="e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb" exitCode=0 Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.079664 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjv54" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.079719 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjv54" event={"ID":"733d7f35-d085-4201-8f2d-d1452f410ad0","Type":"ContainerDied","Data":"e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb"} Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.081696 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjv54" event={"ID":"733d7f35-d085-4201-8f2d-d1452f410ad0","Type":"ContainerDied","Data":"258b7972938ea6674f695896402498771a95465c3b9e4d557b29d6ebf59bed28"} Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.081830 4969 scope.go:117] "RemoveContainer" containerID="e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.117041 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjv54"] Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.120600 4969 scope.go:117] "RemoveContainer" containerID="c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.126365 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjv54"] Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.143459 4969 scope.go:117] "RemoveContainer" containerID="84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.217907 4969 scope.go:117] "RemoveContainer" containerID="e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb" Oct 04 09:24:12 crc kubenswrapper[4969]: E1004 09:24:12.218326 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb\": container with ID starting with e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb not found: ID does not exist" containerID="e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.218356 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb"} err="failed to get container status \"e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb\": rpc error: code = NotFound desc = could not find container \"e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb\": container with ID starting with e33340390020f473305421862ce37856926c5505c763310c6bc22bb965ad3efb not found: ID does not exist" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.218377 4969 scope.go:117] "RemoveContainer" containerID="c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386" Oct 04 09:24:12 crc kubenswrapper[4969]: E1004 09:24:12.218662 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386\": container with ID starting with c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386 not found: ID does not exist" containerID="c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.218694 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386"} err="failed to get container status \"c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386\": rpc error: code = NotFound desc = could not find container \"c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386\": container with ID starting with c47399cb15491d899e080539b71b8a86bdc3d6380a05f1aa3e616f0fa5ba1386 not found: ID does not exist" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.218707 4969 scope.go:117] "RemoveContainer" containerID="84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d" Oct 04 09:24:12 crc kubenswrapper[4969]: E1004 09:24:12.218984 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d\": container with ID starting with 84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d not found: ID does not exist" containerID="84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d" Oct 04 09:24:12 crc kubenswrapper[4969]: I1004 09:24:12.219004 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d"} err="failed to get container status \"84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d\": rpc error: code = NotFound desc = could not find container \"84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d\": container with ID starting with 84a621666fdbee7d2f7fc9730a46c2f62b755a76c1e05cb852a66fdea135ce0d not found: ID does not exist" Oct 04 09:24:13 crc kubenswrapper[4969]: I1004 09:24:13.077912 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="733d7f35-d085-4201-8f2d-d1452f410ad0" path="/var/lib/kubelet/pods/733d7f35-d085-4201-8f2d-d1452f410ad0/volumes" Oct 04 09:26:19 crc kubenswrapper[4969]: I1004 09:26:19.666446 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:26:19 crc kubenswrapper[4969]: I1004 09:26:19.667008 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:26:49 crc kubenswrapper[4969]: I1004 09:26:49.666526 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:26:49 crc kubenswrapper[4969]: I1004 09:26:49.667192 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:27:19 crc kubenswrapper[4969]: I1004 09:27:19.666730 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:27:19 crc kubenswrapper[4969]: I1004 09:27:19.667689 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:27:19 crc kubenswrapper[4969]: I1004 09:27:19.667801 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:27:19 crc kubenswrapper[4969]: I1004 09:27:19.669139 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"baaf5b20707ce5ede797eb8602534733af5d13077a0e17e8fdf8e0c669c98383"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:27:19 crc kubenswrapper[4969]: I1004 09:27:19.669282 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://baaf5b20707ce5ede797eb8602534733af5d13077a0e17e8fdf8e0c669c98383" gracePeriod=600 Oct 04 09:27:20 crc kubenswrapper[4969]: I1004 09:27:20.327175 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="baaf5b20707ce5ede797eb8602534733af5d13077a0e17e8fdf8e0c669c98383" exitCode=0 Oct 04 09:27:20 crc kubenswrapper[4969]: I1004 09:27:20.327276 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"baaf5b20707ce5ede797eb8602534733af5d13077a0e17e8fdf8e0c669c98383"} Oct 04 09:27:20 crc kubenswrapper[4969]: I1004 09:27:20.327585 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05"} Oct 04 09:27:20 crc kubenswrapper[4969]: I1004 09:27:20.327602 4969 scope.go:117] "RemoveContainer" containerID="781b434697fa6da2863c96c6648afe0319c4214de8fe88d9fb17eacb39d7dc29" Oct 04 09:29:49 crc kubenswrapper[4969]: I1004 09:29:49.666036 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:29:49 crc kubenswrapper[4969]: I1004 09:29:49.667342 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.175773 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc"] Oct 04 09:30:00 crc kubenswrapper[4969]: E1004 09:30:00.177368 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerName="extract-content" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.177402 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerName="extract-content" Oct 04 09:30:00 crc kubenswrapper[4969]: E1004 09:30:00.177501 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerName="extract-utilities" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.177524 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerName="extract-utilities" Oct 04 09:30:00 crc kubenswrapper[4969]: E1004 09:30:00.177552 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerName="registry-server" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.177571 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerName="registry-server" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.178042 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="733d7f35-d085-4201-8f2d-d1452f410ad0" containerName="registry-server" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.179517 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.182635 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.183311 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.194398 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc"] Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.203799 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfblb\" (UniqueName: \"kubernetes.io/projected/da10623d-a97e-4a3b-8021-9adb17eae705-kube-api-access-pfblb\") pod \"collect-profiles-29326170-ngplc\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.203849 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da10623d-a97e-4a3b-8021-9adb17eae705-config-volume\") pod \"collect-profiles-29326170-ngplc\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.203963 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da10623d-a97e-4a3b-8021-9adb17eae705-secret-volume\") pod \"collect-profiles-29326170-ngplc\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.305747 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da10623d-a97e-4a3b-8021-9adb17eae705-secret-volume\") pod \"collect-profiles-29326170-ngplc\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.306268 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfblb\" (UniqueName: \"kubernetes.io/projected/da10623d-a97e-4a3b-8021-9adb17eae705-kube-api-access-pfblb\") pod \"collect-profiles-29326170-ngplc\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.306324 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da10623d-a97e-4a3b-8021-9adb17eae705-config-volume\") pod \"collect-profiles-29326170-ngplc\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.307854 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da10623d-a97e-4a3b-8021-9adb17eae705-config-volume\") pod \"collect-profiles-29326170-ngplc\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.312379 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da10623d-a97e-4a3b-8021-9adb17eae705-secret-volume\") pod \"collect-profiles-29326170-ngplc\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.323734 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfblb\" (UniqueName: \"kubernetes.io/projected/da10623d-a97e-4a3b-8021-9adb17eae705-kube-api-access-pfblb\") pod \"collect-profiles-29326170-ngplc\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:00 crc kubenswrapper[4969]: I1004 09:30:00.513499 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:01 crc kubenswrapper[4969]: I1004 09:30:01.029058 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc"] Oct 04 09:30:01 crc kubenswrapper[4969]: I1004 09:30:01.321655 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" event={"ID":"da10623d-a97e-4a3b-8021-9adb17eae705","Type":"ContainerStarted","Data":"72c2f9e0c9ad047df56637c80b8db06e120f20a65e6081ceab5c8ca18b252a48"} Oct 04 09:30:02 crc kubenswrapper[4969]: I1004 09:30:02.332067 4969 generic.go:334] "Generic (PLEG): container finished" podID="da10623d-a97e-4a3b-8021-9adb17eae705" containerID="aedd781c96767ffc0cf343a8c0f2f59df2189d80b62d408394cfd96d77c5c73a" exitCode=0 Oct 04 09:30:02 crc kubenswrapper[4969]: I1004 09:30:02.332175 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" event={"ID":"da10623d-a97e-4a3b-8021-9adb17eae705","Type":"ContainerDied","Data":"aedd781c96767ffc0cf343a8c0f2f59df2189d80b62d408394cfd96d77c5c73a"} Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.354968 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" event={"ID":"da10623d-a97e-4a3b-8021-9adb17eae705","Type":"ContainerDied","Data":"72c2f9e0c9ad047df56637c80b8db06e120f20a65e6081ceab5c8ca18b252a48"} Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.355302 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72c2f9e0c9ad047df56637c80b8db06e120f20a65e6081ceab5c8ca18b252a48" Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.394957 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.491780 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da10623d-a97e-4a3b-8021-9adb17eae705-config-volume\") pod \"da10623d-a97e-4a3b-8021-9adb17eae705\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.492121 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfblb\" (UniqueName: \"kubernetes.io/projected/da10623d-a97e-4a3b-8021-9adb17eae705-kube-api-access-pfblb\") pod \"da10623d-a97e-4a3b-8021-9adb17eae705\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.492261 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da10623d-a97e-4a3b-8021-9adb17eae705-secret-volume\") pod \"da10623d-a97e-4a3b-8021-9adb17eae705\" (UID: \"da10623d-a97e-4a3b-8021-9adb17eae705\") " Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.492677 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da10623d-a97e-4a3b-8021-9adb17eae705-config-volume" (OuterVolumeSpecName: "config-volume") pod "da10623d-a97e-4a3b-8021-9adb17eae705" (UID: "da10623d-a97e-4a3b-8021-9adb17eae705"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.493502 4969 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da10623d-a97e-4a3b-8021-9adb17eae705-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.518769 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da10623d-a97e-4a3b-8021-9adb17eae705-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "da10623d-a97e-4a3b-8021-9adb17eae705" (UID: "da10623d-a97e-4a3b-8021-9adb17eae705"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.519039 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da10623d-a97e-4a3b-8021-9adb17eae705-kube-api-access-pfblb" (OuterVolumeSpecName: "kube-api-access-pfblb") pod "da10623d-a97e-4a3b-8021-9adb17eae705" (UID: "da10623d-a97e-4a3b-8021-9adb17eae705"). InnerVolumeSpecName "kube-api-access-pfblb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.595746 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfblb\" (UniqueName: \"kubernetes.io/projected/da10623d-a97e-4a3b-8021-9adb17eae705-kube-api-access-pfblb\") on node \"crc\" DevicePath \"\"" Oct 04 09:30:04 crc kubenswrapper[4969]: I1004 09:30:04.595970 4969 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da10623d-a97e-4a3b-8021-9adb17eae705-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 09:30:05 crc kubenswrapper[4969]: I1004 09:30:05.364841 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326170-ngplc" Oct 04 09:30:05 crc kubenswrapper[4969]: I1004 09:30:05.496264 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8"] Oct 04 09:30:05 crc kubenswrapper[4969]: I1004 09:30:05.506583 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326125-58fk8"] Oct 04 09:30:07 crc kubenswrapper[4969]: I1004 09:30:07.077488 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69cdc4bb-3b8e-4094-9ee6-fab85d869ffa" path="/var/lib/kubelet/pods/69cdc4bb-3b8e-4094-9ee6-fab85d869ffa/volumes" Oct 04 09:30:19 crc kubenswrapper[4969]: I1004 09:30:19.675042 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:30:19 crc kubenswrapper[4969]: I1004 09:30:19.675815 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:30:33 crc kubenswrapper[4969]: I1004 09:30:33.981203 4969 scope.go:117] "RemoveContainer" containerID="f61d3630bbda83374faa3d0d7b73f1eedd62c00d83d055900270d976c485026d" Oct 04 09:30:49 crc kubenswrapper[4969]: I1004 09:30:49.666119 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:30:49 crc kubenswrapper[4969]: I1004 09:30:49.666746 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:30:49 crc kubenswrapper[4969]: I1004 09:30:49.666810 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:30:49 crc kubenswrapper[4969]: I1004 09:30:49.667907 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:30:49 crc kubenswrapper[4969]: I1004 09:30:49.667997 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" gracePeriod=600 Oct 04 09:30:49 crc kubenswrapper[4969]: E1004 09:30:49.700782 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69d1f843_03d6_403f_8ab3_796e2c97af4f.slice/crio-conmon-971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05.scope\": RecentStats: unable to find data in memory cache]" Oct 04 09:30:49 crc kubenswrapper[4969]: E1004 09:30:49.800002 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:30:49 crc kubenswrapper[4969]: I1004 09:30:49.887700 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" exitCode=0 Oct 04 09:30:49 crc kubenswrapper[4969]: I1004 09:30:49.887769 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05"} Oct 04 09:30:49 crc kubenswrapper[4969]: I1004 09:30:49.887818 4969 scope.go:117] "RemoveContainer" containerID="baaf5b20707ce5ede797eb8602534733af5d13077a0e17e8fdf8e0c669c98383" Oct 04 09:30:49 crc kubenswrapper[4969]: I1004 09:30:49.888852 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:30:49 crc kubenswrapper[4969]: E1004 09:30:49.889499 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:31:00 crc kubenswrapper[4969]: I1004 09:31:00.055927 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:31:00 crc kubenswrapper[4969]: E1004 09:31:00.056718 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:31:07 crc kubenswrapper[4969]: I1004 09:31:07.951038 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f4vg9"] Oct 04 09:31:07 crc kubenswrapper[4969]: E1004 09:31:07.952198 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da10623d-a97e-4a3b-8021-9adb17eae705" containerName="collect-profiles" Oct 04 09:31:07 crc kubenswrapper[4969]: I1004 09:31:07.952215 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="da10623d-a97e-4a3b-8021-9adb17eae705" containerName="collect-profiles" Oct 04 09:31:07 crc kubenswrapper[4969]: I1004 09:31:07.952500 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="da10623d-a97e-4a3b-8021-9adb17eae705" containerName="collect-profiles" Oct 04 09:31:07 crc kubenswrapper[4969]: I1004 09:31:07.954267 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:07 crc kubenswrapper[4969]: I1004 09:31:07.960755 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f4vg9"] Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.000265 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-catalog-content\") pod \"certified-operators-f4vg9\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.000552 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-utilities\") pod \"certified-operators-f4vg9\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.000987 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvrnx\" (UniqueName: \"kubernetes.io/projected/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-kube-api-access-zvrnx\") pod \"certified-operators-f4vg9\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.103115 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-utilities\") pod \"certified-operators-f4vg9\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.103253 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvrnx\" (UniqueName: \"kubernetes.io/projected/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-kube-api-access-zvrnx\") pod \"certified-operators-f4vg9\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.103300 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-catalog-content\") pod \"certified-operators-f4vg9\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.103764 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-catalog-content\") pod \"certified-operators-f4vg9\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.104595 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-utilities\") pod \"certified-operators-f4vg9\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.143346 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvrnx\" (UniqueName: \"kubernetes.io/projected/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-kube-api-access-zvrnx\") pod \"certified-operators-f4vg9\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.293534 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:08 crc kubenswrapper[4969]: I1004 09:31:08.826173 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f4vg9"] Oct 04 09:31:09 crc kubenswrapper[4969]: I1004 09:31:09.144930 4969 generic.go:334] "Generic (PLEG): container finished" podID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerID="721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124" exitCode=0 Oct 04 09:31:09 crc kubenswrapper[4969]: I1004 09:31:09.145007 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4vg9" event={"ID":"4847f54a-2eda-43ac-9db2-9f5fc2c539cb","Type":"ContainerDied","Data":"721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124"} Oct 04 09:31:09 crc kubenswrapper[4969]: I1004 09:31:09.145077 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4vg9" event={"ID":"4847f54a-2eda-43ac-9db2-9f5fc2c539cb","Type":"ContainerStarted","Data":"3b640d751bc57ab9487aca31b907819259eef746fba06f43234d24caaad6138d"} Oct 04 09:31:09 crc kubenswrapper[4969]: I1004 09:31:09.148692 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 09:31:11 crc kubenswrapper[4969]: I1004 09:31:11.176898 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4vg9" event={"ID":"4847f54a-2eda-43ac-9db2-9f5fc2c539cb","Type":"ContainerStarted","Data":"396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a"} Oct 04 09:31:13 crc kubenswrapper[4969]: I1004 09:31:13.086700 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:31:13 crc kubenswrapper[4969]: E1004 09:31:13.087550 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:31:13 crc kubenswrapper[4969]: I1004 09:31:13.204677 4969 generic.go:334] "Generic (PLEG): container finished" podID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerID="396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a" exitCode=0 Oct 04 09:31:13 crc kubenswrapper[4969]: I1004 09:31:13.204756 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4vg9" event={"ID":"4847f54a-2eda-43ac-9db2-9f5fc2c539cb","Type":"ContainerDied","Data":"396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a"} Oct 04 09:31:14 crc kubenswrapper[4969]: I1004 09:31:14.216645 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4vg9" event={"ID":"4847f54a-2eda-43ac-9db2-9f5fc2c539cb","Type":"ContainerStarted","Data":"c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9"} Oct 04 09:31:14 crc kubenswrapper[4969]: I1004 09:31:14.245614 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f4vg9" podStartSLOduration=2.529949967 podStartE2EDuration="7.245588165s" podCreationTimestamp="2025-10-04 09:31:07 +0000 UTC" firstStartedPulling="2025-10-04 09:31:09.148332056 +0000 UTC m=+4496.902600880" lastFinishedPulling="2025-10-04 09:31:13.863970254 +0000 UTC m=+4501.618239078" observedRunningTime="2025-10-04 09:31:14.236594551 +0000 UTC m=+4501.990863365" watchObservedRunningTime="2025-10-04 09:31:14.245588165 +0000 UTC m=+4501.999856999" Oct 04 09:31:18 crc kubenswrapper[4969]: I1004 09:31:18.294458 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:18 crc kubenswrapper[4969]: I1004 09:31:18.294872 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:18 crc kubenswrapper[4969]: I1004 09:31:18.366237 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:19 crc kubenswrapper[4969]: I1004 09:31:19.408367 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:19 crc kubenswrapper[4969]: I1004 09:31:19.472664 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f4vg9"] Oct 04 09:31:21 crc kubenswrapper[4969]: I1004 09:31:21.334253 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f4vg9" podUID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerName="registry-server" containerID="cri-o://c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9" gracePeriod=2 Oct 04 09:31:21 crc kubenswrapper[4969]: I1004 09:31:21.972198 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.112438 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-catalog-content\") pod \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.112483 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-utilities\") pod \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.112668 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvrnx\" (UniqueName: \"kubernetes.io/projected/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-kube-api-access-zvrnx\") pod \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\" (UID: \"4847f54a-2eda-43ac-9db2-9f5fc2c539cb\") " Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.114107 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-utilities" (OuterVolumeSpecName: "utilities") pod "4847f54a-2eda-43ac-9db2-9f5fc2c539cb" (UID: "4847f54a-2eda-43ac-9db2-9f5fc2c539cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.121865 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-kube-api-access-zvrnx" (OuterVolumeSpecName: "kube-api-access-zvrnx") pod "4847f54a-2eda-43ac-9db2-9f5fc2c539cb" (UID: "4847f54a-2eda-43ac-9db2-9f5fc2c539cb"). InnerVolumeSpecName "kube-api-access-zvrnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.169632 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4847f54a-2eda-43ac-9db2-9f5fc2c539cb" (UID: "4847f54a-2eda-43ac-9db2-9f5fc2c539cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.216002 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.216048 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.216067 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvrnx\" (UniqueName: \"kubernetes.io/projected/4847f54a-2eda-43ac-9db2-9f5fc2c539cb-kube-api-access-zvrnx\") on node \"crc\" DevicePath \"\"" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.346983 4969 generic.go:334] "Generic (PLEG): container finished" podID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerID="c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9" exitCode=0 Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.347065 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4vg9" event={"ID":"4847f54a-2eda-43ac-9db2-9f5fc2c539cb","Type":"ContainerDied","Data":"c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9"} Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.347219 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4vg9" event={"ID":"4847f54a-2eda-43ac-9db2-9f5fc2c539cb","Type":"ContainerDied","Data":"3b640d751bc57ab9487aca31b907819259eef746fba06f43234d24caaad6138d"} Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.347242 4969 scope.go:117] "RemoveContainer" containerID="c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.347094 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f4vg9" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.382936 4969 scope.go:117] "RemoveContainer" containerID="396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.401292 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f4vg9"] Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.413378 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f4vg9"] Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.414638 4969 scope.go:117] "RemoveContainer" containerID="721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.478547 4969 scope.go:117] "RemoveContainer" containerID="c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9" Oct 04 09:31:22 crc kubenswrapper[4969]: E1004 09:31:22.479010 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9\": container with ID starting with c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9 not found: ID does not exist" containerID="c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.479049 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9"} err="failed to get container status \"c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9\": rpc error: code = NotFound desc = could not find container \"c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9\": container with ID starting with c6e8f610341af8effdcd3d84f63870ee1eee70d10368d12c6343b4397cca3af9 not found: ID does not exist" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.479075 4969 scope.go:117] "RemoveContainer" containerID="396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a" Oct 04 09:31:22 crc kubenswrapper[4969]: E1004 09:31:22.479329 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a\": container with ID starting with 396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a not found: ID does not exist" containerID="396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.479361 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a"} err="failed to get container status \"396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a\": rpc error: code = NotFound desc = could not find container \"396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a\": container with ID starting with 396ed9d00cdc60afb503f9538f7adca22a990f39939df5354b311a1ca9e8577a not found: ID does not exist" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.479378 4969 scope.go:117] "RemoveContainer" containerID="721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124" Oct 04 09:31:22 crc kubenswrapper[4969]: E1004 09:31:22.479868 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124\": container with ID starting with 721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124 not found: ID does not exist" containerID="721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124" Oct 04 09:31:22 crc kubenswrapper[4969]: I1004 09:31:22.480046 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124"} err="failed to get container status \"721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124\": rpc error: code = NotFound desc = could not find container \"721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124\": container with ID starting with 721eff3bca5e9e2500852556aeb26579d6008aa13b3016d6e7a3cb5238973124 not found: ID does not exist" Oct 04 09:31:23 crc kubenswrapper[4969]: I1004 09:31:23.077195 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" path="/var/lib/kubelet/pods/4847f54a-2eda-43ac-9db2-9f5fc2c539cb/volumes" Oct 04 09:31:27 crc kubenswrapper[4969]: I1004 09:31:27.055480 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:31:27 crc kubenswrapper[4969]: E1004 09:31:27.056557 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:31:40 crc kubenswrapper[4969]: I1004 09:31:40.056471 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:31:40 crc kubenswrapper[4969]: E1004 09:31:40.057393 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:31:55 crc kubenswrapper[4969]: I1004 09:31:55.056510 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:31:55 crc kubenswrapper[4969]: E1004 09:31:55.057921 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:32:09 crc kubenswrapper[4969]: I1004 09:32:09.056501 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:32:09 crc kubenswrapper[4969]: E1004 09:32:09.057695 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:32:23 crc kubenswrapper[4969]: I1004 09:32:23.070149 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:32:23 crc kubenswrapper[4969]: E1004 09:32:23.071577 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:32:38 crc kubenswrapper[4969]: I1004 09:32:38.057335 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:32:38 crc kubenswrapper[4969]: E1004 09:32:38.058462 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:32:52 crc kubenswrapper[4969]: I1004 09:32:52.055697 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:32:52 crc kubenswrapper[4969]: E1004 09:32:52.057085 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:33:04 crc kubenswrapper[4969]: I1004 09:33:04.055859 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:33:04 crc kubenswrapper[4969]: E1004 09:33:04.056679 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:33:18 crc kubenswrapper[4969]: I1004 09:33:18.056302 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:33:18 crc kubenswrapper[4969]: E1004 09:33:18.057548 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:33:30 crc kubenswrapper[4969]: I1004 09:33:30.055848 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:33:30 crc kubenswrapper[4969]: E1004 09:33:30.058154 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:33:43 crc kubenswrapper[4969]: I1004 09:33:43.061265 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:33:43 crc kubenswrapper[4969]: E1004 09:33:43.061903 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:33:49 crc kubenswrapper[4969]: I1004 09:33:49.977163 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p8dq4"] Oct 04 09:33:49 crc kubenswrapper[4969]: E1004 09:33:49.978366 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerName="registry-server" Oct 04 09:33:49 crc kubenswrapper[4969]: I1004 09:33:49.978389 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerName="registry-server" Oct 04 09:33:49 crc kubenswrapper[4969]: E1004 09:33:49.978518 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerName="extract-utilities" Oct 04 09:33:49 crc kubenswrapper[4969]: I1004 09:33:49.978534 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerName="extract-utilities" Oct 04 09:33:49 crc kubenswrapper[4969]: E1004 09:33:49.978560 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerName="extract-content" Oct 04 09:33:49 crc kubenswrapper[4969]: I1004 09:33:49.978572 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerName="extract-content" Oct 04 09:33:49 crc kubenswrapper[4969]: I1004 09:33:49.978901 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="4847f54a-2eda-43ac-9db2-9f5fc2c539cb" containerName="registry-server" Oct 04 09:33:49 crc kubenswrapper[4969]: I1004 09:33:49.981522 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.009806 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p8dq4"] Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.058158 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgjx8\" (UniqueName: \"kubernetes.io/projected/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-kube-api-access-lgjx8\") pod \"community-operators-p8dq4\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.058206 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-catalog-content\") pod \"community-operators-p8dq4\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.058247 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-utilities\") pod \"community-operators-p8dq4\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.160632 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgjx8\" (UniqueName: \"kubernetes.io/projected/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-kube-api-access-lgjx8\") pod \"community-operators-p8dq4\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.160677 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-catalog-content\") pod \"community-operators-p8dq4\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.160699 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-utilities\") pod \"community-operators-p8dq4\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.161487 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-utilities\") pod \"community-operators-p8dq4\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.161711 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-catalog-content\") pod \"community-operators-p8dq4\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.180416 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgjx8\" (UniqueName: \"kubernetes.io/projected/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-kube-api-access-lgjx8\") pod \"community-operators-p8dq4\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.346893 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:33:50 crc kubenswrapper[4969]: I1004 09:33:50.880857 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p8dq4"] Oct 04 09:33:51 crc kubenswrapper[4969]: I1004 09:33:51.080011 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8dq4" event={"ID":"c28a2e09-cc0e-45c7-8d82-452fe1ea432c","Type":"ContainerStarted","Data":"eb7e4c58e03f00748ae59889f50ec0b54479fe0368f12654d376f687fa814f08"} Oct 04 09:33:52 crc kubenswrapper[4969]: I1004 09:33:52.094900 4969 generic.go:334] "Generic (PLEG): container finished" podID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerID="855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85" exitCode=0 Oct 04 09:33:52 crc kubenswrapper[4969]: I1004 09:33:52.094945 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8dq4" event={"ID":"c28a2e09-cc0e-45c7-8d82-452fe1ea432c","Type":"ContainerDied","Data":"855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85"} Oct 04 09:33:53 crc kubenswrapper[4969]: I1004 09:33:53.134119 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8dq4" event={"ID":"c28a2e09-cc0e-45c7-8d82-452fe1ea432c","Type":"ContainerStarted","Data":"57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c"} Oct 04 09:33:54 crc kubenswrapper[4969]: I1004 09:33:54.150995 4969 generic.go:334] "Generic (PLEG): container finished" podID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerID="57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c" exitCode=0 Oct 04 09:33:54 crc kubenswrapper[4969]: I1004 09:33:54.151067 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8dq4" event={"ID":"c28a2e09-cc0e-45c7-8d82-452fe1ea432c","Type":"ContainerDied","Data":"57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c"} Oct 04 09:33:55 crc kubenswrapper[4969]: I1004 09:33:55.167459 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8dq4" event={"ID":"c28a2e09-cc0e-45c7-8d82-452fe1ea432c","Type":"ContainerStarted","Data":"3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124"} Oct 04 09:33:55 crc kubenswrapper[4969]: I1004 09:33:55.201238 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p8dq4" podStartSLOduration=3.6977851790000003 podStartE2EDuration="6.201216427s" podCreationTimestamp="2025-10-04 09:33:49 +0000 UTC" firstStartedPulling="2025-10-04 09:33:52.098542968 +0000 UTC m=+4659.852811822" lastFinishedPulling="2025-10-04 09:33:54.601974226 +0000 UTC m=+4662.356243070" observedRunningTime="2025-10-04 09:33:55.197127436 +0000 UTC m=+4662.951396290" watchObservedRunningTime="2025-10-04 09:33:55.201216427 +0000 UTC m=+4662.955485251" Oct 04 09:33:56 crc kubenswrapper[4969]: I1004 09:33:56.055712 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:33:56 crc kubenswrapper[4969]: E1004 09:33:56.056027 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:34:00 crc kubenswrapper[4969]: I1004 09:34:00.348120 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:34:00 crc kubenswrapper[4969]: I1004 09:34:00.349642 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:34:00 crc kubenswrapper[4969]: I1004 09:34:00.409691 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:34:01 crc kubenswrapper[4969]: I1004 09:34:01.337108 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:34:01 crc kubenswrapper[4969]: I1004 09:34:01.406558 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p8dq4"] Oct 04 09:34:03 crc kubenswrapper[4969]: I1004 09:34:03.268329 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p8dq4" podUID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerName="registry-server" containerID="cri-o://3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124" gracePeriod=2 Oct 04 09:34:03 crc kubenswrapper[4969]: I1004 09:34:03.795805 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:34:03 crc kubenswrapper[4969]: I1004 09:34:03.950288 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-utilities\") pod \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " Oct 04 09:34:03 crc kubenswrapper[4969]: I1004 09:34:03.950642 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-catalog-content\") pod \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " Oct 04 09:34:03 crc kubenswrapper[4969]: I1004 09:34:03.950759 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgjx8\" (UniqueName: \"kubernetes.io/projected/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-kube-api-access-lgjx8\") pod \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\" (UID: \"c28a2e09-cc0e-45c7-8d82-452fe1ea432c\") " Oct 04 09:34:03 crc kubenswrapper[4969]: I1004 09:34:03.951208 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-utilities" (OuterVolumeSpecName: "utilities") pod "c28a2e09-cc0e-45c7-8d82-452fe1ea432c" (UID: "c28a2e09-cc0e-45c7-8d82-452fe1ea432c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:34:03 crc kubenswrapper[4969]: I1004 09:34:03.951543 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:34:03 crc kubenswrapper[4969]: I1004 09:34:03.960371 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-kube-api-access-lgjx8" (OuterVolumeSpecName: "kube-api-access-lgjx8") pod "c28a2e09-cc0e-45c7-8d82-452fe1ea432c" (UID: "c28a2e09-cc0e-45c7-8d82-452fe1ea432c"). InnerVolumeSpecName "kube-api-access-lgjx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.016888 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c28a2e09-cc0e-45c7-8d82-452fe1ea432c" (UID: "c28a2e09-cc0e-45c7-8d82-452fe1ea432c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.053598 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.053641 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgjx8\" (UniqueName: \"kubernetes.io/projected/c28a2e09-cc0e-45c7-8d82-452fe1ea432c-kube-api-access-lgjx8\") on node \"crc\" DevicePath \"\"" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.278946 4969 generic.go:334] "Generic (PLEG): container finished" podID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerID="3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124" exitCode=0 Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.278989 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8dq4" event={"ID":"c28a2e09-cc0e-45c7-8d82-452fe1ea432c","Type":"ContainerDied","Data":"3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124"} Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.279004 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8dq4" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.279017 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8dq4" event={"ID":"c28a2e09-cc0e-45c7-8d82-452fe1ea432c","Type":"ContainerDied","Data":"eb7e4c58e03f00748ae59889f50ec0b54479fe0368f12654d376f687fa814f08"} Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.279038 4969 scope.go:117] "RemoveContainer" containerID="3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.334109 4969 scope.go:117] "RemoveContainer" containerID="57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.341746 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p8dq4"] Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.360695 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p8dq4"] Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.364129 4969 scope.go:117] "RemoveContainer" containerID="855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.407333 4969 scope.go:117] "RemoveContainer" containerID="3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124" Oct 04 09:34:04 crc kubenswrapper[4969]: E1004 09:34:04.407928 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124\": container with ID starting with 3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124 not found: ID does not exist" containerID="3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.407960 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124"} err="failed to get container status \"3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124\": rpc error: code = NotFound desc = could not find container \"3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124\": container with ID starting with 3194c11eacdde6279e4cf67270e13edaff2421c2a9dffcefc5ea6076a56e1124 not found: ID does not exist" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.407983 4969 scope.go:117] "RemoveContainer" containerID="57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c" Oct 04 09:34:04 crc kubenswrapper[4969]: E1004 09:34:04.408367 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c\": container with ID starting with 57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c not found: ID does not exist" containerID="57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.408443 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c"} err="failed to get container status \"57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c\": rpc error: code = NotFound desc = could not find container \"57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c\": container with ID starting with 57a63ee2d387b684de0123d0297b32db8ca41568ed54936c8e26e80945a9693c not found: ID does not exist" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.408489 4969 scope.go:117] "RemoveContainer" containerID="855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85" Oct 04 09:34:04 crc kubenswrapper[4969]: E1004 09:34:04.408824 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85\": container with ID starting with 855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85 not found: ID does not exist" containerID="855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85" Oct 04 09:34:04 crc kubenswrapper[4969]: I1004 09:34:04.408872 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85"} err="failed to get container status \"855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85\": rpc error: code = NotFound desc = could not find container \"855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85\": container with ID starting with 855c1f2c68be87a7dcf2a3b7ed025d2b2318b845ca7448a9a91eda32b51fca85 not found: ID does not exist" Oct 04 09:34:05 crc kubenswrapper[4969]: I1004 09:34:05.075611 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" path="/var/lib/kubelet/pods/c28a2e09-cc0e-45c7-8d82-452fe1ea432c/volumes" Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.795354 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bq77f"] Oct 04 09:34:07 crc kubenswrapper[4969]: E1004 09:34:07.796414 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerName="extract-content" Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.796447 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerName="extract-content" Oct 04 09:34:07 crc kubenswrapper[4969]: E1004 09:34:07.796493 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerName="extract-utilities" Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.796503 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerName="extract-utilities" Oct 04 09:34:07 crc kubenswrapper[4969]: E1004 09:34:07.796531 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerName="registry-server" Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.796539 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerName="registry-server" Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.796794 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c28a2e09-cc0e-45c7-8d82-452fe1ea432c" containerName="registry-server" Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.798608 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.805340 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bq77f"] Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.931221 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-utilities\") pod \"redhat-operators-bq77f\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.931328 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-catalog-content\") pod \"redhat-operators-bq77f\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:07 crc kubenswrapper[4969]: I1004 09:34:07.931350 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htl45\" (UniqueName: \"kubernetes.io/projected/61164321-ad30-49d7-8673-399f5bc66b69-kube-api-access-htl45\") pod \"redhat-operators-bq77f\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:08 crc kubenswrapper[4969]: I1004 09:34:08.032578 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-utilities\") pod \"redhat-operators-bq77f\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:08 crc kubenswrapper[4969]: I1004 09:34:08.032696 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-catalog-content\") pod \"redhat-operators-bq77f\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:08 crc kubenswrapper[4969]: I1004 09:34:08.032719 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htl45\" (UniqueName: \"kubernetes.io/projected/61164321-ad30-49d7-8673-399f5bc66b69-kube-api-access-htl45\") pod \"redhat-operators-bq77f\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:08 crc kubenswrapper[4969]: I1004 09:34:08.033056 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-utilities\") pod \"redhat-operators-bq77f\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:08 crc kubenswrapper[4969]: I1004 09:34:08.033286 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-catalog-content\") pod \"redhat-operators-bq77f\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:08 crc kubenswrapper[4969]: I1004 09:34:08.055650 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:34:08 crc kubenswrapper[4969]: E1004 09:34:08.055994 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:34:08 crc kubenswrapper[4969]: I1004 09:34:08.148050 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htl45\" (UniqueName: \"kubernetes.io/projected/61164321-ad30-49d7-8673-399f5bc66b69-kube-api-access-htl45\") pod \"redhat-operators-bq77f\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:08 crc kubenswrapper[4969]: I1004 09:34:08.420083 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:08 crc kubenswrapper[4969]: I1004 09:34:08.934176 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bq77f"] Oct 04 09:34:08 crc kubenswrapper[4969]: W1004 09:34:08.941962 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61164321_ad30_49d7_8673_399f5bc66b69.slice/crio-cfb896bad23623f7357a61f7c9b7da700b92f25b093f27b952bb9ded31428585 WatchSource:0}: Error finding container cfb896bad23623f7357a61f7c9b7da700b92f25b093f27b952bb9ded31428585: Status 404 returned error can't find the container with id cfb896bad23623f7357a61f7c9b7da700b92f25b093f27b952bb9ded31428585 Oct 04 09:34:09 crc kubenswrapper[4969]: I1004 09:34:09.336295 4969 generic.go:334] "Generic (PLEG): container finished" podID="61164321-ad30-49d7-8673-399f5bc66b69" containerID="d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd" exitCode=0 Oct 04 09:34:09 crc kubenswrapper[4969]: I1004 09:34:09.336379 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bq77f" event={"ID":"61164321-ad30-49d7-8673-399f5bc66b69","Type":"ContainerDied","Data":"d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd"} Oct 04 09:34:09 crc kubenswrapper[4969]: I1004 09:34:09.336444 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bq77f" event={"ID":"61164321-ad30-49d7-8673-399f5bc66b69","Type":"ContainerStarted","Data":"cfb896bad23623f7357a61f7c9b7da700b92f25b093f27b952bb9ded31428585"} Oct 04 09:34:10 crc kubenswrapper[4969]: I1004 09:34:10.345316 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bq77f" event={"ID":"61164321-ad30-49d7-8673-399f5bc66b69","Type":"ContainerStarted","Data":"9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4"} Oct 04 09:34:14 crc kubenswrapper[4969]: I1004 09:34:14.395937 4969 generic.go:334] "Generic (PLEG): container finished" podID="61164321-ad30-49d7-8673-399f5bc66b69" containerID="9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4" exitCode=0 Oct 04 09:34:14 crc kubenswrapper[4969]: I1004 09:34:14.396077 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bq77f" event={"ID":"61164321-ad30-49d7-8673-399f5bc66b69","Type":"ContainerDied","Data":"9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4"} Oct 04 09:34:16 crc kubenswrapper[4969]: I1004 09:34:16.431877 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bq77f" event={"ID":"61164321-ad30-49d7-8673-399f5bc66b69","Type":"ContainerStarted","Data":"435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5"} Oct 04 09:34:16 crc kubenswrapper[4969]: I1004 09:34:16.458219 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bq77f" podStartSLOduration=3.5573574040000002 podStartE2EDuration="9.458189994s" podCreationTimestamp="2025-10-04 09:34:07 +0000 UTC" firstStartedPulling="2025-10-04 09:34:09.339310965 +0000 UTC m=+4677.093579769" lastFinishedPulling="2025-10-04 09:34:15.240143535 +0000 UTC m=+4682.994412359" observedRunningTime="2025-10-04 09:34:16.457062185 +0000 UTC m=+4684.211331009" watchObservedRunningTime="2025-10-04 09:34:16.458189994 +0000 UTC m=+4684.212458838" Oct 04 09:34:18 crc kubenswrapper[4969]: I1004 09:34:18.421626 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:18 crc kubenswrapper[4969]: I1004 09:34:18.421912 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:19 crc kubenswrapper[4969]: I1004 09:34:19.510648 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bq77f" podUID="61164321-ad30-49d7-8673-399f5bc66b69" containerName="registry-server" probeResult="failure" output=< Oct 04 09:34:19 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:34:19 crc kubenswrapper[4969]: > Oct 04 09:34:21 crc kubenswrapper[4969]: I1004 09:34:21.055693 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:34:21 crc kubenswrapper[4969]: E1004 09:34:21.056064 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:34:28 crc kubenswrapper[4969]: I1004 09:34:28.506236 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:28 crc kubenswrapper[4969]: I1004 09:34:28.577648 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:28 crc kubenswrapper[4969]: I1004 09:34:28.760803 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bq77f"] Oct 04 09:34:29 crc kubenswrapper[4969]: I1004 09:34:29.597174 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bq77f" podUID="61164321-ad30-49d7-8673-399f5bc66b69" containerName="registry-server" containerID="cri-o://435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5" gracePeriod=2 Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.177108 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.230814 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htl45\" (UniqueName: \"kubernetes.io/projected/61164321-ad30-49d7-8673-399f5bc66b69-kube-api-access-htl45\") pod \"61164321-ad30-49d7-8673-399f5bc66b69\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.230874 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-catalog-content\") pod \"61164321-ad30-49d7-8673-399f5bc66b69\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.230897 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-utilities\") pod \"61164321-ad30-49d7-8673-399f5bc66b69\" (UID: \"61164321-ad30-49d7-8673-399f5bc66b69\") " Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.232467 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-utilities" (OuterVolumeSpecName: "utilities") pod "61164321-ad30-49d7-8673-399f5bc66b69" (UID: "61164321-ad30-49d7-8673-399f5bc66b69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.244779 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61164321-ad30-49d7-8673-399f5bc66b69-kube-api-access-htl45" (OuterVolumeSpecName: "kube-api-access-htl45") pod "61164321-ad30-49d7-8673-399f5bc66b69" (UID: "61164321-ad30-49d7-8673-399f5bc66b69"). InnerVolumeSpecName "kube-api-access-htl45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.326640 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61164321-ad30-49d7-8673-399f5bc66b69" (UID: "61164321-ad30-49d7-8673-399f5bc66b69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.334576 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htl45\" (UniqueName: \"kubernetes.io/projected/61164321-ad30-49d7-8673-399f5bc66b69-kube-api-access-htl45\") on node \"crc\" DevicePath \"\"" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.334650 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.334663 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61164321-ad30-49d7-8673-399f5bc66b69-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.613885 4969 generic.go:334] "Generic (PLEG): container finished" podID="61164321-ad30-49d7-8673-399f5bc66b69" containerID="435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5" exitCode=0 Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.613945 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bq77f" event={"ID":"61164321-ad30-49d7-8673-399f5bc66b69","Type":"ContainerDied","Data":"435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5"} Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.614698 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bq77f" event={"ID":"61164321-ad30-49d7-8673-399f5bc66b69","Type":"ContainerDied","Data":"cfb896bad23623f7357a61f7c9b7da700b92f25b093f27b952bb9ded31428585"} Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.614761 4969 scope.go:117] "RemoveContainer" containerID="435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.614005 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bq77f" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.654256 4969 scope.go:117] "RemoveContainer" containerID="9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.656172 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bq77f"] Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.678016 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bq77f"] Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.688334 4969 scope.go:117] "RemoveContainer" containerID="d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.735772 4969 scope.go:117] "RemoveContainer" containerID="435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5" Oct 04 09:34:30 crc kubenswrapper[4969]: E1004 09:34:30.736555 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5\": container with ID starting with 435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5 not found: ID does not exist" containerID="435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.736677 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5"} err="failed to get container status \"435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5\": rpc error: code = NotFound desc = could not find container \"435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5\": container with ID starting with 435787e75e8c91583a7e8443d871fcf4320492497e2ee6182c3d91061c40b0c5 not found: ID does not exist" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.736790 4969 scope.go:117] "RemoveContainer" containerID="9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4" Oct 04 09:34:30 crc kubenswrapper[4969]: E1004 09:34:30.737216 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4\": container with ID starting with 9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4 not found: ID does not exist" containerID="9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.737327 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4"} err="failed to get container status \"9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4\": rpc error: code = NotFound desc = could not find container \"9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4\": container with ID starting with 9a43493db7af0923dd8d84a49843d106e8f1ba402ae7db5322c2620360a41af4 not found: ID does not exist" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.737444 4969 scope.go:117] "RemoveContainer" containerID="d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd" Oct 04 09:34:30 crc kubenswrapper[4969]: E1004 09:34:30.737918 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd\": container with ID starting with d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd not found: ID does not exist" containerID="d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd" Oct 04 09:34:30 crc kubenswrapper[4969]: I1004 09:34:30.738024 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd"} err="failed to get container status \"d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd\": rpc error: code = NotFound desc = could not find container \"d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd\": container with ID starting with d733f52a2b0abb749df2798b40e3dbb9b26608f708e438636332287c853b87cd not found: ID does not exist" Oct 04 09:34:31 crc kubenswrapper[4969]: I1004 09:34:31.082127 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61164321-ad30-49d7-8673-399f5bc66b69" path="/var/lib/kubelet/pods/61164321-ad30-49d7-8673-399f5bc66b69/volumes" Oct 04 09:34:35 crc kubenswrapper[4969]: I1004 09:34:35.055295 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:34:35 crc kubenswrapper[4969]: E1004 09:34:35.056103 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:34:50 crc kubenswrapper[4969]: I1004 09:34:50.055331 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:34:50 crc kubenswrapper[4969]: E1004 09:34:50.056247 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:35:05 crc kubenswrapper[4969]: I1004 09:35:05.056135 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:35:05 crc kubenswrapper[4969]: E1004 09:35:05.057238 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:35:16 crc kubenswrapper[4969]: I1004 09:35:16.055808 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:35:16 crc kubenswrapper[4969]: E1004 09:35:16.056880 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.649739 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lmjn4"] Oct 04 09:35:17 crc kubenswrapper[4969]: E1004 09:35:17.650968 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61164321-ad30-49d7-8673-399f5bc66b69" containerName="extract-content" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.651039 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="61164321-ad30-49d7-8673-399f5bc66b69" containerName="extract-content" Oct 04 09:35:17 crc kubenswrapper[4969]: E1004 09:35:17.651108 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61164321-ad30-49d7-8673-399f5bc66b69" containerName="extract-utilities" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.651159 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="61164321-ad30-49d7-8673-399f5bc66b69" containerName="extract-utilities" Oct 04 09:35:17 crc kubenswrapper[4969]: E1004 09:35:17.651232 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61164321-ad30-49d7-8673-399f5bc66b69" containerName="registry-server" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.651284 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="61164321-ad30-49d7-8673-399f5bc66b69" containerName="registry-server" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.651569 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="61164321-ad30-49d7-8673-399f5bc66b69" containerName="registry-server" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.653201 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.668919 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmjn4"] Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.800373 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-utilities\") pod \"redhat-marketplace-lmjn4\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.800440 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx2md\" (UniqueName: \"kubernetes.io/projected/295e9243-7b93-4485-a56d-449998f0dd50-kube-api-access-cx2md\") pod \"redhat-marketplace-lmjn4\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.801273 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-catalog-content\") pod \"redhat-marketplace-lmjn4\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.903049 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-utilities\") pod \"redhat-marketplace-lmjn4\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.903105 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx2md\" (UniqueName: \"kubernetes.io/projected/295e9243-7b93-4485-a56d-449998f0dd50-kube-api-access-cx2md\") pod \"redhat-marketplace-lmjn4\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.903260 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-catalog-content\") pod \"redhat-marketplace-lmjn4\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.903946 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-catalog-content\") pod \"redhat-marketplace-lmjn4\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.903989 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-utilities\") pod \"redhat-marketplace-lmjn4\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.928410 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx2md\" (UniqueName: \"kubernetes.io/projected/295e9243-7b93-4485-a56d-449998f0dd50-kube-api-access-cx2md\") pod \"redhat-marketplace-lmjn4\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:17 crc kubenswrapper[4969]: I1004 09:35:17.989714 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:18 crc kubenswrapper[4969]: I1004 09:35:18.485525 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmjn4"] Oct 04 09:35:18 crc kubenswrapper[4969]: W1004 09:35:18.495319 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod295e9243_7b93_4485_a56d_449998f0dd50.slice/crio-bf5962699d27adbb13416b5851321a560eb7c2613f72296d18eef6b24390e627 WatchSource:0}: Error finding container bf5962699d27adbb13416b5851321a560eb7c2613f72296d18eef6b24390e627: Status 404 returned error can't find the container with id bf5962699d27adbb13416b5851321a560eb7c2613f72296d18eef6b24390e627 Oct 04 09:35:19 crc kubenswrapper[4969]: I1004 09:35:19.169469 4969 generic.go:334] "Generic (PLEG): container finished" podID="295e9243-7b93-4485-a56d-449998f0dd50" containerID="7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6" exitCode=0 Oct 04 09:35:19 crc kubenswrapper[4969]: I1004 09:35:19.169541 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmjn4" event={"ID":"295e9243-7b93-4485-a56d-449998f0dd50","Type":"ContainerDied","Data":"7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6"} Oct 04 09:35:19 crc kubenswrapper[4969]: I1004 09:35:19.169584 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmjn4" event={"ID":"295e9243-7b93-4485-a56d-449998f0dd50","Type":"ContainerStarted","Data":"bf5962699d27adbb13416b5851321a560eb7c2613f72296d18eef6b24390e627"} Oct 04 09:35:21 crc kubenswrapper[4969]: I1004 09:35:21.199604 4969 generic.go:334] "Generic (PLEG): container finished" podID="295e9243-7b93-4485-a56d-449998f0dd50" containerID="659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007" exitCode=0 Oct 04 09:35:21 crc kubenswrapper[4969]: I1004 09:35:21.199734 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmjn4" event={"ID":"295e9243-7b93-4485-a56d-449998f0dd50","Type":"ContainerDied","Data":"659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007"} Oct 04 09:35:22 crc kubenswrapper[4969]: I1004 09:35:22.219884 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmjn4" event={"ID":"295e9243-7b93-4485-a56d-449998f0dd50","Type":"ContainerStarted","Data":"834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd"} Oct 04 09:35:22 crc kubenswrapper[4969]: I1004 09:35:22.246045 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lmjn4" podStartSLOduration=2.774662046 podStartE2EDuration="5.246019748s" podCreationTimestamp="2025-10-04 09:35:17 +0000 UTC" firstStartedPulling="2025-10-04 09:35:19.17293707 +0000 UTC m=+4746.927205924" lastFinishedPulling="2025-10-04 09:35:21.644294782 +0000 UTC m=+4749.398563626" observedRunningTime="2025-10-04 09:35:22.241831102 +0000 UTC m=+4749.996099986" watchObservedRunningTime="2025-10-04 09:35:22.246019748 +0000 UTC m=+4750.000288602" Oct 04 09:35:27 crc kubenswrapper[4969]: I1004 09:35:27.990912 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:27 crc kubenswrapper[4969]: I1004 09:35:27.991954 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:28 crc kubenswrapper[4969]: I1004 09:35:28.228355 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:28 crc kubenswrapper[4969]: I1004 09:35:28.390528 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:28 crc kubenswrapper[4969]: I1004 09:35:28.480356 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmjn4"] Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.055946 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:35:30 crc kubenswrapper[4969]: E1004 09:35:30.056548 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.333964 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lmjn4" podUID="295e9243-7b93-4485-a56d-449998f0dd50" containerName="registry-server" containerID="cri-o://834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd" gracePeriod=2 Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.837868 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.895890 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-utilities\") pod \"295e9243-7b93-4485-a56d-449998f0dd50\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.895953 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-catalog-content\") pod \"295e9243-7b93-4485-a56d-449998f0dd50\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.896100 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx2md\" (UniqueName: \"kubernetes.io/projected/295e9243-7b93-4485-a56d-449998f0dd50-kube-api-access-cx2md\") pod \"295e9243-7b93-4485-a56d-449998f0dd50\" (UID: \"295e9243-7b93-4485-a56d-449998f0dd50\") " Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.898965 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-utilities" (OuterVolumeSpecName: "utilities") pod "295e9243-7b93-4485-a56d-449998f0dd50" (UID: "295e9243-7b93-4485-a56d-449998f0dd50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.902559 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/295e9243-7b93-4485-a56d-449998f0dd50-kube-api-access-cx2md" (OuterVolumeSpecName: "kube-api-access-cx2md") pod "295e9243-7b93-4485-a56d-449998f0dd50" (UID: "295e9243-7b93-4485-a56d-449998f0dd50"). InnerVolumeSpecName "kube-api-access-cx2md". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.913626 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "295e9243-7b93-4485-a56d-449998f0dd50" (UID: "295e9243-7b93-4485-a56d-449998f0dd50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.999417 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx2md\" (UniqueName: \"kubernetes.io/projected/295e9243-7b93-4485-a56d-449998f0dd50-kube-api-access-cx2md\") on node \"crc\" DevicePath \"\"" Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.999493 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:35:30 crc kubenswrapper[4969]: I1004 09:35:30.999512 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295e9243-7b93-4485-a56d-449998f0dd50-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.351235 4969 generic.go:334] "Generic (PLEG): container finished" podID="295e9243-7b93-4485-a56d-449998f0dd50" containerID="834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd" exitCode=0 Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.351314 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmjn4" event={"ID":"295e9243-7b93-4485-a56d-449998f0dd50","Type":"ContainerDied","Data":"834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd"} Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.351657 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmjn4" event={"ID":"295e9243-7b93-4485-a56d-449998f0dd50","Type":"ContainerDied","Data":"bf5962699d27adbb13416b5851321a560eb7c2613f72296d18eef6b24390e627"} Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.351339 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lmjn4" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.351685 4969 scope.go:117] "RemoveContainer" containerID="834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.391695 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmjn4"] Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.401894 4969 scope.go:117] "RemoveContainer" containerID="659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.410145 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmjn4"] Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.444458 4969 scope.go:117] "RemoveContainer" containerID="7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.505527 4969 scope.go:117] "RemoveContainer" containerID="834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd" Oct 04 09:35:31 crc kubenswrapper[4969]: E1004 09:35:31.506323 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd\": container with ID starting with 834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd not found: ID does not exist" containerID="834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.506403 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd"} err="failed to get container status \"834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd\": rpc error: code = NotFound desc = could not find container \"834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd\": container with ID starting with 834558b1721e279ea0c9709c8f4d18e0c8fe3c69697fdf1b83bc683ae263aebd not found: ID does not exist" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.506475 4969 scope.go:117] "RemoveContainer" containerID="659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007" Oct 04 09:35:31 crc kubenswrapper[4969]: E1004 09:35:31.507058 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007\": container with ID starting with 659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007 not found: ID does not exist" containerID="659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.507114 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007"} err="failed to get container status \"659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007\": rpc error: code = NotFound desc = could not find container \"659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007\": container with ID starting with 659ee0894ee6f79cdd293bfab035dae7759c8ac2241f827ba2e35b71d5d1b007 not found: ID does not exist" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.507150 4969 scope.go:117] "RemoveContainer" containerID="7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6" Oct 04 09:35:31 crc kubenswrapper[4969]: E1004 09:35:31.507636 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6\": container with ID starting with 7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6 not found: ID does not exist" containerID="7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6" Oct 04 09:35:31 crc kubenswrapper[4969]: I1004 09:35:31.507748 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6"} err="failed to get container status \"7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6\": rpc error: code = NotFound desc = could not find container \"7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6\": container with ID starting with 7808fa50e8a7dc02bc3bfcfe054b5fc4438d7f99b05d1597970f6a94ee2d47e6 not found: ID does not exist" Oct 04 09:35:33 crc kubenswrapper[4969]: I1004 09:35:33.078801 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="295e9243-7b93-4485-a56d-449998f0dd50" path="/var/lib/kubelet/pods/295e9243-7b93-4485-a56d-449998f0dd50/volumes" Oct 04 09:35:45 crc kubenswrapper[4969]: I1004 09:35:45.055801 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:35:45 crc kubenswrapper[4969]: E1004 09:35:45.057246 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:36:00 crc kubenswrapper[4969]: I1004 09:36:00.056014 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:36:00 crc kubenswrapper[4969]: I1004 09:36:00.697620 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"f91d2a1bd2aa39e16057048d8b8a77371a8867ef4bf7ebb93aba4a4257c97ab0"} Oct 04 09:38:19 crc kubenswrapper[4969]: I1004 09:38:19.666472 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:38:19 crc kubenswrapper[4969]: I1004 09:38:19.667150 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:38:49 crc kubenswrapper[4969]: I1004 09:38:49.666649 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:38:49 crc kubenswrapper[4969]: I1004 09:38:49.667257 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:39:19 crc kubenswrapper[4969]: I1004 09:39:19.666959 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:39:19 crc kubenswrapper[4969]: I1004 09:39:19.667814 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:39:19 crc kubenswrapper[4969]: I1004 09:39:19.667911 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:39:19 crc kubenswrapper[4969]: I1004 09:39:19.668919 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f91d2a1bd2aa39e16057048d8b8a77371a8867ef4bf7ebb93aba4a4257c97ab0"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:39:19 crc kubenswrapper[4969]: I1004 09:39:19.668999 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://f91d2a1bd2aa39e16057048d8b8a77371a8867ef4bf7ebb93aba4a4257c97ab0" gracePeriod=600 Oct 04 09:39:20 crc kubenswrapper[4969]: I1004 09:39:20.056321 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="f91d2a1bd2aa39e16057048d8b8a77371a8867ef4bf7ebb93aba4a4257c97ab0" exitCode=0 Oct 04 09:39:20 crc kubenswrapper[4969]: I1004 09:39:20.056368 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"f91d2a1bd2aa39e16057048d8b8a77371a8867ef4bf7ebb93aba4a4257c97ab0"} Oct 04 09:39:20 crc kubenswrapper[4969]: I1004 09:39:20.057222 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676"} Oct 04 09:39:20 crc kubenswrapper[4969]: I1004 09:39:20.057250 4969 scope.go:117] "RemoveContainer" containerID="971710f9fdb7f592c19c701bef89b1afd6677479c62094015ea45204fe7e2b05" Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.854686 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ph6cn"] Oct 04 09:41:41 crc kubenswrapper[4969]: E1004 09:41:41.856195 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295e9243-7b93-4485-a56d-449998f0dd50" containerName="extract-content" Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.856223 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="295e9243-7b93-4485-a56d-449998f0dd50" containerName="extract-content" Oct 04 09:41:41 crc kubenswrapper[4969]: E1004 09:41:41.856255 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295e9243-7b93-4485-a56d-449998f0dd50" containerName="extract-utilities" Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.856268 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="295e9243-7b93-4485-a56d-449998f0dd50" containerName="extract-utilities" Oct 04 09:41:41 crc kubenswrapper[4969]: E1004 09:41:41.856325 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295e9243-7b93-4485-a56d-449998f0dd50" containerName="registry-server" Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.856339 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="295e9243-7b93-4485-a56d-449998f0dd50" containerName="registry-server" Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.856749 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="295e9243-7b93-4485-a56d-449998f0dd50" containerName="registry-server" Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.859711 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.872828 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ph6cn"] Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.899201 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c8n6\" (UniqueName: \"kubernetes.io/projected/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-kube-api-access-2c8n6\") pod \"certified-operators-ph6cn\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.899371 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-catalog-content\") pod \"certified-operators-ph6cn\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:41 crc kubenswrapper[4969]: I1004 09:41:41.899722 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-utilities\") pod \"certified-operators-ph6cn\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:42 crc kubenswrapper[4969]: I1004 09:41:42.001443 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-catalog-content\") pod \"certified-operators-ph6cn\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:42 crc kubenswrapper[4969]: I1004 09:41:42.001881 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-utilities\") pod \"certified-operators-ph6cn\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:42 crc kubenswrapper[4969]: I1004 09:41:42.002122 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c8n6\" (UniqueName: \"kubernetes.io/projected/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-kube-api-access-2c8n6\") pod \"certified-operators-ph6cn\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:42 crc kubenswrapper[4969]: I1004 09:41:42.002292 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-utilities\") pod \"certified-operators-ph6cn\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:42 crc kubenswrapper[4969]: I1004 09:41:42.002295 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-catalog-content\") pod \"certified-operators-ph6cn\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:42 crc kubenswrapper[4969]: I1004 09:41:42.033671 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c8n6\" (UniqueName: \"kubernetes.io/projected/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-kube-api-access-2c8n6\") pod \"certified-operators-ph6cn\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:42 crc kubenswrapper[4969]: I1004 09:41:42.184626 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:42 crc kubenswrapper[4969]: I1004 09:41:42.677331 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ph6cn"] Oct 04 09:41:43 crc kubenswrapper[4969]: I1004 09:41:43.800666 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ph6cn" event={"ID":"fe30336f-f3dd-4ba8-8c10-7b20ae92098d","Type":"ContainerStarted","Data":"c1203fac788388df8ef0bdfa78fb20a59c91927d31f3d8b0c725154de53feef5"} Oct 04 09:41:44 crc kubenswrapper[4969]: I1004 09:41:44.811257 4969 generic.go:334] "Generic (PLEG): container finished" podID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerID="d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1" exitCode=0 Oct 04 09:41:44 crc kubenswrapper[4969]: I1004 09:41:44.811329 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ph6cn" event={"ID":"fe30336f-f3dd-4ba8-8c10-7b20ae92098d","Type":"ContainerDied","Data":"d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1"} Oct 04 09:41:44 crc kubenswrapper[4969]: I1004 09:41:44.813890 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 09:41:45 crc kubenswrapper[4969]: I1004 09:41:45.825330 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ph6cn" event={"ID":"fe30336f-f3dd-4ba8-8c10-7b20ae92098d","Type":"ContainerStarted","Data":"b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c"} Oct 04 09:41:46 crc kubenswrapper[4969]: I1004 09:41:46.840678 4969 generic.go:334] "Generic (PLEG): container finished" podID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerID="b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c" exitCode=0 Oct 04 09:41:46 crc kubenswrapper[4969]: I1004 09:41:46.840745 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ph6cn" event={"ID":"fe30336f-f3dd-4ba8-8c10-7b20ae92098d","Type":"ContainerDied","Data":"b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c"} Oct 04 09:41:47 crc kubenswrapper[4969]: I1004 09:41:47.863974 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ph6cn" event={"ID":"fe30336f-f3dd-4ba8-8c10-7b20ae92098d","Type":"ContainerStarted","Data":"1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73"} Oct 04 09:41:47 crc kubenswrapper[4969]: I1004 09:41:47.903013 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ph6cn" podStartSLOduration=4.375311051 podStartE2EDuration="6.902996594s" podCreationTimestamp="2025-10-04 09:41:41 +0000 UTC" firstStartedPulling="2025-10-04 09:41:44.813587801 +0000 UTC m=+5132.567856615" lastFinishedPulling="2025-10-04 09:41:47.341273344 +0000 UTC m=+5135.095542158" observedRunningTime="2025-10-04 09:41:47.899581498 +0000 UTC m=+5135.653850312" watchObservedRunningTime="2025-10-04 09:41:47.902996594 +0000 UTC m=+5135.657265408" Oct 04 09:41:49 crc kubenswrapper[4969]: I1004 09:41:49.666553 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:41:49 crc kubenswrapper[4969]: I1004 09:41:49.666813 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:41:52 crc kubenswrapper[4969]: I1004 09:41:52.185201 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:52 crc kubenswrapper[4969]: I1004 09:41:52.185701 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:52 crc kubenswrapper[4969]: I1004 09:41:52.260904 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:52 crc kubenswrapper[4969]: I1004 09:41:52.987246 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:53 crc kubenswrapper[4969]: I1004 09:41:53.042888 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ph6cn"] Oct 04 09:41:54 crc kubenswrapper[4969]: I1004 09:41:54.941791 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ph6cn" podUID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerName="registry-server" containerID="cri-o://1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73" gracePeriod=2 Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.539689 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.696986 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-catalog-content\") pod \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.697035 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-utilities\") pod \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.697265 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c8n6\" (UniqueName: \"kubernetes.io/projected/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-kube-api-access-2c8n6\") pod \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\" (UID: \"fe30336f-f3dd-4ba8-8c10-7b20ae92098d\") " Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.697766 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-utilities" (OuterVolumeSpecName: "utilities") pod "fe30336f-f3dd-4ba8-8c10-7b20ae92098d" (UID: "fe30336f-f3dd-4ba8-8c10-7b20ae92098d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.702710 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-kube-api-access-2c8n6" (OuterVolumeSpecName: "kube-api-access-2c8n6") pod "fe30336f-f3dd-4ba8-8c10-7b20ae92098d" (UID: "fe30336f-f3dd-4ba8-8c10-7b20ae92098d"). InnerVolumeSpecName "kube-api-access-2c8n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.742397 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe30336f-f3dd-4ba8-8c10-7b20ae92098d" (UID: "fe30336f-f3dd-4ba8-8c10-7b20ae92098d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.799900 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c8n6\" (UniqueName: \"kubernetes.io/projected/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-kube-api-access-2c8n6\") on node \"crc\" DevicePath \"\"" Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.799930 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.799939 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe30336f-f3dd-4ba8-8c10-7b20ae92098d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.963196 4969 generic.go:334] "Generic (PLEG): container finished" podID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerID="1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73" exitCode=0 Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.963256 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ph6cn" event={"ID":"fe30336f-f3dd-4ba8-8c10-7b20ae92098d","Type":"ContainerDied","Data":"1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73"} Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.963273 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ph6cn" Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.963306 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ph6cn" event={"ID":"fe30336f-f3dd-4ba8-8c10-7b20ae92098d","Type":"ContainerDied","Data":"c1203fac788388df8ef0bdfa78fb20a59c91927d31f3d8b0c725154de53feef5"} Oct 04 09:41:55 crc kubenswrapper[4969]: I1004 09:41:55.963340 4969 scope.go:117] "RemoveContainer" containerID="1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73" Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.000282 4969 scope.go:117] "RemoveContainer" containerID="b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c" Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.028366 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ph6cn"] Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.042374 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ph6cn"] Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.069918 4969 scope.go:117] "RemoveContainer" containerID="d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1" Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.124640 4969 scope.go:117] "RemoveContainer" containerID="1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73" Oct 04 09:41:56 crc kubenswrapper[4969]: E1004 09:41:56.125246 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73\": container with ID starting with 1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73 not found: ID does not exist" containerID="1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73" Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.125308 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73"} err="failed to get container status \"1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73\": rpc error: code = NotFound desc = could not find container \"1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73\": container with ID starting with 1f2dee06567c122b63684a31c1acf89aa60cbf59117d0b804d81560a5251eb73 not found: ID does not exist" Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.125348 4969 scope.go:117] "RemoveContainer" containerID="b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c" Oct 04 09:41:56 crc kubenswrapper[4969]: E1004 09:41:56.125959 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c\": container with ID starting with b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c not found: ID does not exist" containerID="b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c" Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.125992 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c"} err="failed to get container status \"b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c\": rpc error: code = NotFound desc = could not find container \"b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c\": container with ID starting with b9b95dee47ea647f4bcabb6cdf2fa722d2850269015c42e459ace57f96ac797c not found: ID does not exist" Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.126011 4969 scope.go:117] "RemoveContainer" containerID="d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1" Oct 04 09:41:56 crc kubenswrapper[4969]: E1004 09:41:56.126316 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1\": container with ID starting with d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1 not found: ID does not exist" containerID="d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1" Oct 04 09:41:56 crc kubenswrapper[4969]: I1004 09:41:56.126343 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1"} err="failed to get container status \"d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1\": rpc error: code = NotFound desc = could not find container \"d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1\": container with ID starting with d4f10b907dac0aa473935a32c86c15a36adf8aad5c059e7715b99d56d93e77e1 not found: ID does not exist" Oct 04 09:41:57 crc kubenswrapper[4969]: I1004 09:41:57.085812 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" path="/var/lib/kubelet/pods/fe30336f-f3dd-4ba8-8c10-7b20ae92098d/volumes" Oct 04 09:42:16 crc kubenswrapper[4969]: E1004 09:42:16.771093 4969 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.195:48952->38.102.83.195:43269: write tcp 38.102.83.195:48952->38.102.83.195:43269: write: broken pipe Oct 04 09:42:19 crc kubenswrapper[4969]: I1004 09:42:19.667244 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:42:19 crc kubenswrapper[4969]: I1004 09:42:19.668089 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:42:49 crc kubenswrapper[4969]: I1004 09:42:49.667485 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:42:49 crc kubenswrapper[4969]: I1004 09:42:49.668276 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:42:49 crc kubenswrapper[4969]: I1004 09:42:49.668369 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:42:49 crc kubenswrapper[4969]: I1004 09:42:49.669651 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:42:49 crc kubenswrapper[4969]: I1004 09:42:49.669816 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" gracePeriod=600 Oct 04 09:42:49 crc kubenswrapper[4969]: E1004 09:42:49.823370 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:42:50 crc kubenswrapper[4969]: I1004 09:42:50.656714 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" exitCode=0 Oct 04 09:42:50 crc kubenswrapper[4969]: I1004 09:42:50.656836 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676"} Oct 04 09:42:50 crc kubenswrapper[4969]: I1004 09:42:50.657143 4969 scope.go:117] "RemoveContainer" containerID="f91d2a1bd2aa39e16057048d8b8a77371a8867ef4bf7ebb93aba4a4257c97ab0" Oct 04 09:42:50 crc kubenswrapper[4969]: I1004 09:42:50.658223 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:42:50 crc kubenswrapper[4969]: E1004 09:42:50.658767 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:43:06 crc kubenswrapper[4969]: I1004 09:43:06.055855 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:43:06 crc kubenswrapper[4969]: E1004 09:43:06.057470 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:43:21 crc kubenswrapper[4969]: I1004 09:43:21.056699 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:43:21 crc kubenswrapper[4969]: E1004 09:43:21.057713 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:43:33 crc kubenswrapper[4969]: I1004 09:43:33.061815 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:43:33 crc kubenswrapper[4969]: E1004 09:43:33.062719 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:43:46 crc kubenswrapper[4969]: I1004 09:43:46.056490 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:43:46 crc kubenswrapper[4969]: E1004 09:43:46.057919 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:43:50 crc kubenswrapper[4969]: E1004 09:43:50.171463 4969 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.195:42392->38.102.83.195:43269: write tcp 38.102.83.195:42392->38.102.83.195:43269: write: connection reset by peer Oct 04 09:44:00 crc kubenswrapper[4969]: I1004 09:44:00.056012 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:44:00 crc kubenswrapper[4969]: E1004 09:44:00.056781 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:44:11 crc kubenswrapper[4969]: I1004 09:44:11.055634 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:44:11 crc kubenswrapper[4969]: E1004 09:44:11.056888 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:44:26 crc kubenswrapper[4969]: I1004 09:44:26.055811 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:44:26 crc kubenswrapper[4969]: E1004 09:44:26.056928 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:44:40 crc kubenswrapper[4969]: I1004 09:44:40.060997 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:44:40 crc kubenswrapper[4969]: E1004 09:44:40.062245 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:44:54 crc kubenswrapper[4969]: I1004 09:44:54.055612 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:44:54 crc kubenswrapper[4969]: E1004 09:44:54.056554 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.145566 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp"] Oct 04 09:45:00 crc kubenswrapper[4969]: E1004 09:45:00.146491 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerName="extract-utilities" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.146504 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerName="extract-utilities" Oct 04 09:45:00 crc kubenswrapper[4969]: E1004 09:45:00.146526 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerName="registry-server" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.146534 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerName="registry-server" Oct 04 09:45:00 crc kubenswrapper[4969]: E1004 09:45:00.146549 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerName="extract-content" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.146554 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerName="extract-content" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.146745 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe30336f-f3dd-4ba8-8c10-7b20ae92098d" containerName="registry-server" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.147427 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.151940 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.152181 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.161698 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp"] Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.318679 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l2dk\" (UniqueName: \"kubernetes.io/projected/5d5ac949-655b-4b2c-985d-ac702e75bf35-kube-api-access-6l2dk\") pod \"collect-profiles-29326185-bxhkp\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.319086 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5ac949-655b-4b2c-985d-ac702e75bf35-config-volume\") pod \"collect-profiles-29326185-bxhkp\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.319309 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5ac949-655b-4b2c-985d-ac702e75bf35-secret-volume\") pod \"collect-profiles-29326185-bxhkp\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.419843 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l2dk\" (UniqueName: \"kubernetes.io/projected/5d5ac949-655b-4b2c-985d-ac702e75bf35-kube-api-access-6l2dk\") pod \"collect-profiles-29326185-bxhkp\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.419896 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5ac949-655b-4b2c-985d-ac702e75bf35-config-volume\") pod \"collect-profiles-29326185-bxhkp\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.420014 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5ac949-655b-4b2c-985d-ac702e75bf35-secret-volume\") pod \"collect-profiles-29326185-bxhkp\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.420998 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5ac949-655b-4b2c-985d-ac702e75bf35-config-volume\") pod \"collect-profiles-29326185-bxhkp\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.427974 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5ac949-655b-4b2c-985d-ac702e75bf35-secret-volume\") pod \"collect-profiles-29326185-bxhkp\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.438490 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l2dk\" (UniqueName: \"kubernetes.io/projected/5d5ac949-655b-4b2c-985d-ac702e75bf35-kube-api-access-6l2dk\") pod \"collect-profiles-29326185-bxhkp\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.472886 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:00 crc kubenswrapper[4969]: I1004 09:45:00.930880 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp"] Oct 04 09:45:01 crc kubenswrapper[4969]: I1004 09:45:01.282661 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" event={"ID":"5d5ac949-655b-4b2c-985d-ac702e75bf35","Type":"ContainerStarted","Data":"915f5467a93f243a4abd26f2b2a48ac1cccf2dcdd545b679b150a13acf85649b"} Oct 04 09:45:01 crc kubenswrapper[4969]: I1004 09:45:01.282954 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" event={"ID":"5d5ac949-655b-4b2c-985d-ac702e75bf35","Type":"ContainerStarted","Data":"817beeca61fa1598e608acb554cd642ad375cea2c81ca8ca95ecb0b35bec0b0b"} Oct 04 09:45:01 crc kubenswrapper[4969]: I1004 09:45:01.304400 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" podStartSLOduration=1.304382781 podStartE2EDuration="1.304382781s" podCreationTimestamp="2025-10-04 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 09:45:01.29705248 +0000 UTC m=+5329.051321324" watchObservedRunningTime="2025-10-04 09:45:01.304382781 +0000 UTC m=+5329.058651595" Oct 04 09:45:02 crc kubenswrapper[4969]: I1004 09:45:02.295833 4969 generic.go:334] "Generic (PLEG): container finished" podID="5d5ac949-655b-4b2c-985d-ac702e75bf35" containerID="915f5467a93f243a4abd26f2b2a48ac1cccf2dcdd545b679b150a13acf85649b" exitCode=0 Oct 04 09:45:02 crc kubenswrapper[4969]: I1004 09:45:02.295917 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" event={"ID":"5d5ac949-655b-4b2c-985d-ac702e75bf35","Type":"ContainerDied","Data":"915f5467a93f243a4abd26f2b2a48ac1cccf2dcdd545b679b150a13acf85649b"} Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.753522 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.889548 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l2dk\" (UniqueName: \"kubernetes.io/projected/5d5ac949-655b-4b2c-985d-ac702e75bf35-kube-api-access-6l2dk\") pod \"5d5ac949-655b-4b2c-985d-ac702e75bf35\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.889591 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5ac949-655b-4b2c-985d-ac702e75bf35-config-volume\") pod \"5d5ac949-655b-4b2c-985d-ac702e75bf35\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.889780 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5ac949-655b-4b2c-985d-ac702e75bf35-secret-volume\") pod \"5d5ac949-655b-4b2c-985d-ac702e75bf35\" (UID: \"5d5ac949-655b-4b2c-985d-ac702e75bf35\") " Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.891646 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5ac949-655b-4b2c-985d-ac702e75bf35-config-volume" (OuterVolumeSpecName: "config-volume") pod "5d5ac949-655b-4b2c-985d-ac702e75bf35" (UID: "5d5ac949-655b-4b2c-985d-ac702e75bf35"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.896521 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d5ac949-655b-4b2c-985d-ac702e75bf35-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5d5ac949-655b-4b2c-985d-ac702e75bf35" (UID: "5d5ac949-655b-4b2c-985d-ac702e75bf35"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.896969 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d5ac949-655b-4b2c-985d-ac702e75bf35-kube-api-access-6l2dk" (OuterVolumeSpecName: "kube-api-access-6l2dk") pod "5d5ac949-655b-4b2c-985d-ac702e75bf35" (UID: "5d5ac949-655b-4b2c-985d-ac702e75bf35"). InnerVolumeSpecName "kube-api-access-6l2dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.993015 4969 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5ac949-655b-4b2c-985d-ac702e75bf35-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.993071 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l2dk\" (UniqueName: \"kubernetes.io/projected/5d5ac949-655b-4b2c-985d-ac702e75bf35-kube-api-access-6l2dk\") on node \"crc\" DevicePath \"\"" Oct 04 09:45:03 crc kubenswrapper[4969]: I1004 09:45:03.993080 4969 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5ac949-655b-4b2c-985d-ac702e75bf35-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 09:45:04 crc kubenswrapper[4969]: I1004 09:45:04.321006 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" event={"ID":"5d5ac949-655b-4b2c-985d-ac702e75bf35","Type":"ContainerDied","Data":"817beeca61fa1598e608acb554cd642ad375cea2c81ca8ca95ecb0b35bec0b0b"} Oct 04 09:45:04 crc kubenswrapper[4969]: I1004 09:45:04.321376 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="817beeca61fa1598e608acb554cd642ad375cea2c81ca8ca95ecb0b35bec0b0b" Oct 04 09:45:04 crc kubenswrapper[4969]: I1004 09:45:04.321113 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326185-bxhkp" Oct 04 09:45:04 crc kubenswrapper[4969]: I1004 09:45:04.391566 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52"] Oct 04 09:45:04 crc kubenswrapper[4969]: I1004 09:45:04.409316 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326140-wbs52"] Oct 04 09:45:05 crc kubenswrapper[4969]: I1004 09:45:05.075644 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d8d980d-5a47-45ac-bdc7-d881f07bf40d" path="/var/lib/kubelet/pods/3d8d980d-5a47-45ac-bdc7-d881f07bf40d/volumes" Oct 04 09:45:06 crc kubenswrapper[4969]: I1004 09:45:06.055732 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:45:06 crc kubenswrapper[4969]: E1004 09:45:06.056296 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.794044 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5rflq"] Oct 04 09:45:07 crc kubenswrapper[4969]: E1004 09:45:07.794591 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d5ac949-655b-4b2c-985d-ac702e75bf35" containerName="collect-profiles" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.794606 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d5ac949-655b-4b2c-985d-ac702e75bf35" containerName="collect-profiles" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.794869 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d5ac949-655b-4b2c-985d-ac702e75bf35" containerName="collect-profiles" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.796793 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.803655 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rflq"] Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.877135 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2llx\" (UniqueName: \"kubernetes.io/projected/efffd15d-bc3e-4bfa-849e-ab6d1e856306-kube-api-access-m2llx\") pod \"community-operators-5rflq\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.877218 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-catalog-content\") pod \"community-operators-5rflq\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.877385 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-utilities\") pod \"community-operators-5rflq\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.979547 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2llx\" (UniqueName: \"kubernetes.io/projected/efffd15d-bc3e-4bfa-849e-ab6d1e856306-kube-api-access-m2llx\") pod \"community-operators-5rflq\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.979613 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-catalog-content\") pod \"community-operators-5rflq\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.979716 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-utilities\") pod \"community-operators-5rflq\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.980128 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-catalog-content\") pod \"community-operators-5rflq\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:07 crc kubenswrapper[4969]: I1004 09:45:07.980152 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-utilities\") pod \"community-operators-5rflq\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:08 crc kubenswrapper[4969]: I1004 09:45:08.011852 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2llx\" (UniqueName: \"kubernetes.io/projected/efffd15d-bc3e-4bfa-849e-ab6d1e856306-kube-api-access-m2llx\") pod \"community-operators-5rflq\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:08 crc kubenswrapper[4969]: I1004 09:45:08.134201 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:08 crc kubenswrapper[4969]: I1004 09:45:08.662298 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rflq"] Oct 04 09:45:09 crc kubenswrapper[4969]: I1004 09:45:09.378127 4969 generic.go:334] "Generic (PLEG): container finished" podID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerID="799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a" exitCode=0 Oct 04 09:45:09 crc kubenswrapper[4969]: I1004 09:45:09.378189 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rflq" event={"ID":"efffd15d-bc3e-4bfa-849e-ab6d1e856306","Type":"ContainerDied","Data":"799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a"} Oct 04 09:45:09 crc kubenswrapper[4969]: I1004 09:45:09.378260 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rflq" event={"ID":"efffd15d-bc3e-4bfa-849e-ab6d1e856306","Type":"ContainerStarted","Data":"add4771d9576a35fa8f4f0dcdaf40644c4267418ba2032db6dfe2014a80c58f6"} Oct 04 09:45:10 crc kubenswrapper[4969]: I1004 09:45:10.389187 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rflq" event={"ID":"efffd15d-bc3e-4bfa-849e-ab6d1e856306","Type":"ContainerStarted","Data":"91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6"} Oct 04 09:45:12 crc kubenswrapper[4969]: I1004 09:45:12.409589 4969 generic.go:334] "Generic (PLEG): container finished" podID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerID="91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6" exitCode=0 Oct 04 09:45:12 crc kubenswrapper[4969]: I1004 09:45:12.409666 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rflq" event={"ID":"efffd15d-bc3e-4bfa-849e-ab6d1e856306","Type":"ContainerDied","Data":"91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6"} Oct 04 09:45:13 crc kubenswrapper[4969]: I1004 09:45:13.422196 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rflq" event={"ID":"efffd15d-bc3e-4bfa-849e-ab6d1e856306","Type":"ContainerStarted","Data":"057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d"} Oct 04 09:45:13 crc kubenswrapper[4969]: I1004 09:45:13.477169 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5rflq" podStartSLOduration=3.055312858 podStartE2EDuration="6.477145144s" podCreationTimestamp="2025-10-04 09:45:07 +0000 UTC" firstStartedPulling="2025-10-04 09:45:09.380149219 +0000 UTC m=+5337.134418043" lastFinishedPulling="2025-10-04 09:45:12.801981515 +0000 UTC m=+5340.556250329" observedRunningTime="2025-10-04 09:45:13.459902308 +0000 UTC m=+5341.214171162" watchObservedRunningTime="2025-10-04 09:45:13.477145144 +0000 UTC m=+5341.231413978" Oct 04 09:45:18 crc kubenswrapper[4969]: I1004 09:45:18.055798 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:45:18 crc kubenswrapper[4969]: E1004 09:45:18.056937 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:45:18 crc kubenswrapper[4969]: I1004 09:45:18.148667 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:18 crc kubenswrapper[4969]: I1004 09:45:18.148756 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:18 crc kubenswrapper[4969]: I1004 09:45:18.223072 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:18 crc kubenswrapper[4969]: I1004 09:45:18.600368 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:18 crc kubenswrapper[4969]: I1004 09:45:18.665300 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5rflq"] Oct 04 09:45:20 crc kubenswrapper[4969]: I1004 09:45:20.532485 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5rflq" podUID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerName="registry-server" containerID="cri-o://057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d" gracePeriod=2 Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.025403 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.169725 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2llx\" (UniqueName: \"kubernetes.io/projected/efffd15d-bc3e-4bfa-849e-ab6d1e856306-kube-api-access-m2llx\") pod \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.170940 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-utilities\") pod \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.171015 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-catalog-content\") pod \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\" (UID: \"efffd15d-bc3e-4bfa-849e-ab6d1e856306\") " Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.172380 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-utilities" (OuterVolumeSpecName: "utilities") pod "efffd15d-bc3e-4bfa-849e-ab6d1e856306" (UID: "efffd15d-bc3e-4bfa-849e-ab6d1e856306"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.181306 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efffd15d-bc3e-4bfa-849e-ab6d1e856306-kube-api-access-m2llx" (OuterVolumeSpecName: "kube-api-access-m2llx") pod "efffd15d-bc3e-4bfa-849e-ab6d1e856306" (UID: "efffd15d-bc3e-4bfa-849e-ab6d1e856306"). InnerVolumeSpecName "kube-api-access-m2llx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.229058 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "efffd15d-bc3e-4bfa-849e-ab6d1e856306" (UID: "efffd15d-bc3e-4bfa-849e-ab6d1e856306"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.273408 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2llx\" (UniqueName: \"kubernetes.io/projected/efffd15d-bc3e-4bfa-849e-ab6d1e856306-kube-api-access-m2llx\") on node \"crc\" DevicePath \"\"" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.273465 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.273479 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efffd15d-bc3e-4bfa-849e-ab6d1e856306-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.567236 4969 generic.go:334] "Generic (PLEG): container finished" podID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerID="057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d" exitCode=0 Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.567374 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rflq" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.567460 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rflq" event={"ID":"efffd15d-bc3e-4bfa-849e-ab6d1e856306","Type":"ContainerDied","Data":"057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d"} Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.567786 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rflq" event={"ID":"efffd15d-bc3e-4bfa-849e-ab6d1e856306","Type":"ContainerDied","Data":"add4771d9576a35fa8f4f0dcdaf40644c4267418ba2032db6dfe2014a80c58f6"} Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.567885 4969 scope.go:117] "RemoveContainer" containerID="057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.616613 4969 scope.go:117] "RemoveContainer" containerID="91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.620587 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5rflq"] Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.631594 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5rflq"] Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.649884 4969 scope.go:117] "RemoveContainer" containerID="799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.685655 4969 scope.go:117] "RemoveContainer" containerID="057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d" Oct 04 09:45:21 crc kubenswrapper[4969]: E1004 09:45:21.687051 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d\": container with ID starting with 057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d not found: ID does not exist" containerID="057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.687083 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d"} err="failed to get container status \"057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d\": rpc error: code = NotFound desc = could not find container \"057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d\": container with ID starting with 057ac598687b915eb145a56b67a8db04e8ebe9435d8eec9f5aae3efb0197d71d not found: ID does not exist" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.687103 4969 scope.go:117] "RemoveContainer" containerID="91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6" Oct 04 09:45:21 crc kubenswrapper[4969]: E1004 09:45:21.687734 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6\": container with ID starting with 91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6 not found: ID does not exist" containerID="91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.687777 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6"} err="failed to get container status \"91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6\": rpc error: code = NotFound desc = could not find container \"91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6\": container with ID starting with 91e3dbbdcf368b9893af4d47e484cfbd87e74215f74eb742da0eadea98cb23a6 not found: ID does not exist" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.687793 4969 scope.go:117] "RemoveContainer" containerID="799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a" Oct 04 09:45:21 crc kubenswrapper[4969]: E1004 09:45:21.688028 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a\": container with ID starting with 799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a not found: ID does not exist" containerID="799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a" Oct 04 09:45:21 crc kubenswrapper[4969]: I1004 09:45:21.688048 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a"} err="failed to get container status \"799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a\": rpc error: code = NotFound desc = could not find container \"799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a\": container with ID starting with 799a88e94106f7773e04a452639c4b85d89884075ac3a57b47103f607bcd090a not found: ID does not exist" Oct 04 09:45:23 crc kubenswrapper[4969]: I1004 09:45:23.077552 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" path="/var/lib/kubelet/pods/efffd15d-bc3e-4bfa-849e-ab6d1e856306/volumes" Oct 04 09:45:31 crc kubenswrapper[4969]: I1004 09:45:31.055481 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:45:31 crc kubenswrapper[4969]: E1004 09:45:31.056178 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:45:34 crc kubenswrapper[4969]: I1004 09:45:34.555966 4969 scope.go:117] "RemoveContainer" containerID="12e3cd5901ff41c7811c65879f5f524bf40cd71dfd4579e18e310b7bf9f3b1ee" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.513900 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fdctd"] Oct 04 09:45:39 crc kubenswrapper[4969]: E1004 09:45:39.514866 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerName="extract-content" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.514881 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerName="extract-content" Oct 04 09:45:39 crc kubenswrapper[4969]: E1004 09:45:39.514891 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerName="registry-server" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.514897 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerName="registry-server" Oct 04 09:45:39 crc kubenswrapper[4969]: E1004 09:45:39.514925 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerName="extract-utilities" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.514932 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerName="extract-utilities" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.515129 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="efffd15d-bc3e-4bfa-849e-ab6d1e856306" containerName="registry-server" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.516926 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.526951 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdctd"] Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.577971 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-utilities\") pod \"redhat-marketplace-fdctd\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.578024 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mrqc\" (UniqueName: \"kubernetes.io/projected/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-kube-api-access-8mrqc\") pod \"redhat-marketplace-fdctd\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.578335 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-catalog-content\") pod \"redhat-marketplace-fdctd\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.681156 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-catalog-content\") pod \"redhat-marketplace-fdctd\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.681324 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-utilities\") pod \"redhat-marketplace-fdctd\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.681365 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mrqc\" (UniqueName: \"kubernetes.io/projected/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-kube-api-access-8mrqc\") pod \"redhat-marketplace-fdctd\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.681772 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-utilities\") pod \"redhat-marketplace-fdctd\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.681824 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-catalog-content\") pod \"redhat-marketplace-fdctd\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.710997 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mrqc\" (UniqueName: \"kubernetes.io/projected/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-kube-api-access-8mrqc\") pod \"redhat-marketplace-fdctd\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:39 crc kubenswrapper[4969]: I1004 09:45:39.856283 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:40 crc kubenswrapper[4969]: I1004 09:45:40.332657 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdctd"] Oct 04 09:45:40 crc kubenswrapper[4969]: I1004 09:45:40.794351 4969 generic.go:334] "Generic (PLEG): container finished" podID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerID="dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19" exitCode=0 Oct 04 09:45:40 crc kubenswrapper[4969]: I1004 09:45:40.794681 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdctd" event={"ID":"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048","Type":"ContainerDied","Data":"dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19"} Oct 04 09:45:40 crc kubenswrapper[4969]: I1004 09:45:40.794713 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdctd" event={"ID":"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048","Type":"ContainerStarted","Data":"11a7dd7386f3fdd648c34b8a68eaa3f187cc6dfddcd0a20f26e38cbd9442ec5d"} Oct 04 09:45:41 crc kubenswrapper[4969]: I1004 09:45:41.805161 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdctd" event={"ID":"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048","Type":"ContainerStarted","Data":"405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419"} Oct 04 09:45:42 crc kubenswrapper[4969]: I1004 09:45:42.056368 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:45:42 crc kubenswrapper[4969]: E1004 09:45:42.057062 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:45:42 crc kubenswrapper[4969]: I1004 09:45:42.818621 4969 generic.go:334] "Generic (PLEG): container finished" podID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerID="405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419" exitCode=0 Oct 04 09:45:42 crc kubenswrapper[4969]: I1004 09:45:42.818710 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdctd" event={"ID":"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048","Type":"ContainerDied","Data":"405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419"} Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.303513 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t8djf"] Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.308468 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.327313 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t8djf"] Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.380845 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw9b7\" (UniqueName: \"kubernetes.io/projected/c0c99898-8c47-4801-855a-46f5e2f83669-kube-api-access-hw9b7\") pod \"redhat-operators-t8djf\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.381003 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-utilities\") pod \"redhat-operators-t8djf\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.381135 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-catalog-content\") pod \"redhat-operators-t8djf\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.482846 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-catalog-content\") pod \"redhat-operators-t8djf\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.482966 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw9b7\" (UniqueName: \"kubernetes.io/projected/c0c99898-8c47-4801-855a-46f5e2f83669-kube-api-access-hw9b7\") pod \"redhat-operators-t8djf\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.483066 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-utilities\") pod \"redhat-operators-t8djf\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.483256 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-catalog-content\") pod \"redhat-operators-t8djf\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.483637 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-utilities\") pod \"redhat-operators-t8djf\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.506683 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw9b7\" (UniqueName: \"kubernetes.io/projected/c0c99898-8c47-4801-855a-46f5e2f83669-kube-api-access-hw9b7\") pod \"redhat-operators-t8djf\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.634691 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.837239 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdctd" event={"ID":"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048","Type":"ContainerStarted","Data":"ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd"} Oct 04 09:45:43 crc kubenswrapper[4969]: I1004 09:45:43.879245 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fdctd" podStartSLOduration=2.2943202400000002 podStartE2EDuration="4.879201965s" podCreationTimestamp="2025-10-04 09:45:39 +0000 UTC" firstStartedPulling="2025-10-04 09:45:40.796851306 +0000 UTC m=+5368.551120120" lastFinishedPulling="2025-10-04 09:45:43.381733001 +0000 UTC m=+5371.136001845" observedRunningTime="2025-10-04 09:45:43.873161306 +0000 UTC m=+5371.627430120" watchObservedRunningTime="2025-10-04 09:45:43.879201965 +0000 UTC m=+5371.633470779" Oct 04 09:45:44 crc kubenswrapper[4969]: I1004 09:45:44.170539 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t8djf"] Oct 04 09:45:44 crc kubenswrapper[4969]: I1004 09:45:44.849622 4969 generic.go:334] "Generic (PLEG): container finished" podID="c0c99898-8c47-4801-855a-46f5e2f83669" containerID="0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9" exitCode=0 Oct 04 09:45:44 crc kubenswrapper[4969]: I1004 09:45:44.849720 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8djf" event={"ID":"c0c99898-8c47-4801-855a-46f5e2f83669","Type":"ContainerDied","Data":"0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9"} Oct 04 09:45:44 crc kubenswrapper[4969]: I1004 09:45:44.850274 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8djf" event={"ID":"c0c99898-8c47-4801-855a-46f5e2f83669","Type":"ContainerStarted","Data":"2ebf44bf20aa30c1d619eb4d85a4e7ec0152772e9627522a8f9ddb3090fb14b0"} Oct 04 09:45:46 crc kubenswrapper[4969]: I1004 09:45:46.885352 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8djf" event={"ID":"c0c99898-8c47-4801-855a-46f5e2f83669","Type":"ContainerStarted","Data":"f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b"} Oct 04 09:45:48 crc kubenswrapper[4969]: I1004 09:45:48.906524 4969 generic.go:334] "Generic (PLEG): container finished" podID="c0c99898-8c47-4801-855a-46f5e2f83669" containerID="f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b" exitCode=0 Oct 04 09:45:48 crc kubenswrapper[4969]: I1004 09:45:48.906594 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8djf" event={"ID":"c0c99898-8c47-4801-855a-46f5e2f83669","Type":"ContainerDied","Data":"f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b"} Oct 04 09:45:49 crc kubenswrapper[4969]: I1004 09:45:49.856457 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:49 crc kubenswrapper[4969]: I1004 09:45:49.856846 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:49 crc kubenswrapper[4969]: I1004 09:45:49.917926 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8djf" event={"ID":"c0c99898-8c47-4801-855a-46f5e2f83669","Type":"ContainerStarted","Data":"a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811"} Oct 04 09:45:49 crc kubenswrapper[4969]: I1004 09:45:49.922308 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:49 crc kubenswrapper[4969]: I1004 09:45:49.938557 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t8djf" podStartSLOduration=2.510017631 podStartE2EDuration="6.938538528s" podCreationTimestamp="2025-10-04 09:45:43 +0000 UTC" firstStartedPulling="2025-10-04 09:45:44.852093458 +0000 UTC m=+5372.606362312" lastFinishedPulling="2025-10-04 09:45:49.280614395 +0000 UTC m=+5377.034883209" observedRunningTime="2025-10-04 09:45:49.934367495 +0000 UTC m=+5377.688636309" watchObservedRunningTime="2025-10-04 09:45:49.938538528 +0000 UTC m=+5377.692807352" Oct 04 09:45:50 crc kubenswrapper[4969]: I1004 09:45:50.037728 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:53 crc kubenswrapper[4969]: I1004 09:45:53.635845 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:53 crc kubenswrapper[4969]: I1004 09:45:53.636336 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.092523 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdctd"] Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.092848 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fdctd" podUID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerName="registry-server" containerID="cri-o://ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd" gracePeriod=2 Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.599891 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.711948 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mrqc\" (UniqueName: \"kubernetes.io/projected/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-kube-api-access-8mrqc\") pod \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.712204 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-utilities\") pod \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.712415 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-catalog-content\") pod \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\" (UID: \"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048\") " Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.712685 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-utilities" (OuterVolumeSpecName: "utilities") pod "d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" (UID: "d0a4ae07-3f0f-47e9-ae40-8ecc79d08048"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.713037 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.721046 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-kube-api-access-8mrqc" (OuterVolumeSpecName: "kube-api-access-8mrqc") pod "d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" (UID: "d0a4ae07-3f0f-47e9-ae40-8ecc79d08048"). InnerVolumeSpecName "kube-api-access-8mrqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.724876 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" (UID: "d0a4ae07-3f0f-47e9-ae40-8ecc79d08048"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.728189 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t8djf" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" containerName="registry-server" probeResult="failure" output=< Oct 04 09:45:54 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:45:54 crc kubenswrapper[4969]: > Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.814800 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.814832 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mrqc\" (UniqueName: \"kubernetes.io/projected/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048-kube-api-access-8mrqc\") on node \"crc\" DevicePath \"\"" Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.976832 4969 generic.go:334] "Generic (PLEG): container finished" podID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerID="ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd" exitCode=0 Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.976881 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdctd" Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.976890 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdctd" event={"ID":"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048","Type":"ContainerDied","Data":"ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd"} Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.976967 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdctd" event={"ID":"d0a4ae07-3f0f-47e9-ae40-8ecc79d08048","Type":"ContainerDied","Data":"11a7dd7386f3fdd648c34b8a68eaa3f187cc6dfddcd0a20f26e38cbd9442ec5d"} Oct 04 09:45:54 crc kubenswrapper[4969]: I1004 09:45:54.977001 4969 scope.go:117] "RemoveContainer" containerID="ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.004193 4969 scope.go:117] "RemoveContainer" containerID="405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.015700 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdctd"] Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.029012 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdctd"] Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.051861 4969 scope.go:117] "RemoveContainer" containerID="dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.059201 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:45:55 crc kubenswrapper[4969]: E1004 09:45:55.059541 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.071320 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" path="/var/lib/kubelet/pods/d0a4ae07-3f0f-47e9-ae40-8ecc79d08048/volumes" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.093334 4969 scope.go:117] "RemoveContainer" containerID="ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd" Oct 04 09:45:55 crc kubenswrapper[4969]: E1004 09:45:55.093788 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd\": container with ID starting with ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd not found: ID does not exist" containerID="ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.093858 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd"} err="failed to get container status \"ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd\": rpc error: code = NotFound desc = could not find container \"ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd\": container with ID starting with ee01df511cc80ce31bbb25667af31eb8dd7e0ae073ae79b12e526ddf8c28c0fd not found: ID does not exist" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.093898 4969 scope.go:117] "RemoveContainer" containerID="405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419" Oct 04 09:45:55 crc kubenswrapper[4969]: E1004 09:45:55.094309 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419\": container with ID starting with 405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419 not found: ID does not exist" containerID="405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.094347 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419"} err="failed to get container status \"405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419\": rpc error: code = NotFound desc = could not find container \"405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419\": container with ID starting with 405fa016428cdf949ddab4a51b0b9abec45399c0ce909917b522b1b54a383419 not found: ID does not exist" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.094377 4969 scope.go:117] "RemoveContainer" containerID="dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19" Oct 04 09:45:55 crc kubenswrapper[4969]: E1004 09:45:55.094698 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19\": container with ID starting with dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19 not found: ID does not exist" containerID="dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19" Oct 04 09:45:55 crc kubenswrapper[4969]: I1004 09:45:55.094723 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19"} err="failed to get container status \"dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19\": rpc error: code = NotFound desc = could not find container \"dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19\": container with ID starting with dae5dd47befdc6930fc0f37ba770e80ba52347b7c9d7eaa1b69986c2da6c3e19 not found: ID does not exist" Oct 04 09:46:03 crc kubenswrapper[4969]: I1004 09:46:03.700497 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:46:03 crc kubenswrapper[4969]: I1004 09:46:03.769974 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:46:03 crc kubenswrapper[4969]: I1004 09:46:03.939075 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t8djf"] Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.144522 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t8djf" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" containerName="registry-server" containerID="cri-o://a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811" gracePeriod=2 Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.636829 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.771680 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-catalog-content\") pod \"c0c99898-8c47-4801-855a-46f5e2f83669\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.771797 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-utilities\") pod \"c0c99898-8c47-4801-855a-46f5e2f83669\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.771904 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw9b7\" (UniqueName: \"kubernetes.io/projected/c0c99898-8c47-4801-855a-46f5e2f83669-kube-api-access-hw9b7\") pod \"c0c99898-8c47-4801-855a-46f5e2f83669\" (UID: \"c0c99898-8c47-4801-855a-46f5e2f83669\") " Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.772822 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-utilities" (OuterVolumeSpecName: "utilities") pod "c0c99898-8c47-4801-855a-46f5e2f83669" (UID: "c0c99898-8c47-4801-855a-46f5e2f83669"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.777437 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0c99898-8c47-4801-855a-46f5e2f83669-kube-api-access-hw9b7" (OuterVolumeSpecName: "kube-api-access-hw9b7") pod "c0c99898-8c47-4801-855a-46f5e2f83669" (UID: "c0c99898-8c47-4801-855a-46f5e2f83669"). InnerVolumeSpecName "kube-api-access-hw9b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.851956 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0c99898-8c47-4801-855a-46f5e2f83669" (UID: "c0c99898-8c47-4801-855a-46f5e2f83669"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.874035 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw9b7\" (UniqueName: \"kubernetes.io/projected/c0c99898-8c47-4801-855a-46f5e2f83669-kube-api-access-hw9b7\") on node \"crc\" DevicePath \"\"" Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.874074 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:46:05 crc kubenswrapper[4969]: I1004 09:46:05.874088 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0c99898-8c47-4801-855a-46f5e2f83669-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.160011 4969 generic.go:334] "Generic (PLEG): container finished" podID="c0c99898-8c47-4801-855a-46f5e2f83669" containerID="a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811" exitCode=0 Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.160107 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8djf" event={"ID":"c0c99898-8c47-4801-855a-46f5e2f83669","Type":"ContainerDied","Data":"a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811"} Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.160151 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8djf" event={"ID":"c0c99898-8c47-4801-855a-46f5e2f83669","Type":"ContainerDied","Data":"2ebf44bf20aa30c1d619eb4d85a4e7ec0152772e9627522a8f9ddb3090fb14b0"} Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.160186 4969 scope.go:117] "RemoveContainer" containerID="a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.160404 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8djf" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.188566 4969 scope.go:117] "RemoveContainer" containerID="f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.226852 4969 scope.go:117] "RemoveContainer" containerID="0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.227841 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t8djf"] Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.242145 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t8djf"] Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.273844 4969 scope.go:117] "RemoveContainer" containerID="a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811" Oct 04 09:46:06 crc kubenswrapper[4969]: E1004 09:46:06.274499 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811\": container with ID starting with a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811 not found: ID does not exist" containerID="a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.274565 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811"} err="failed to get container status \"a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811\": rpc error: code = NotFound desc = could not find container \"a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811\": container with ID starting with a7cf71e56712108651ab924f4fabf60ce9d8e4e1742f38521ef1c3c441335811 not found: ID does not exist" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.274606 4969 scope.go:117] "RemoveContainer" containerID="f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b" Oct 04 09:46:06 crc kubenswrapper[4969]: E1004 09:46:06.275063 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b\": container with ID starting with f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b not found: ID does not exist" containerID="f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.275095 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b"} err="failed to get container status \"f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b\": rpc error: code = NotFound desc = could not find container \"f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b\": container with ID starting with f77e59a946d66cda3c8698b1acaa4331b325bcdd1ee8a8f922e380668ee51c4b not found: ID does not exist" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.275120 4969 scope.go:117] "RemoveContainer" containerID="0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9" Oct 04 09:46:06 crc kubenswrapper[4969]: E1004 09:46:06.275770 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9\": container with ID starting with 0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9 not found: ID does not exist" containerID="0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9" Oct 04 09:46:06 crc kubenswrapper[4969]: I1004 09:46:06.275825 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9"} err="failed to get container status \"0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9\": rpc error: code = NotFound desc = could not find container \"0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9\": container with ID starting with 0c191b20d93682b5f96d56710d90aa0e5fc0b671cb7d05396ddc22c8fd0df7f9 not found: ID does not exist" Oct 04 09:46:07 crc kubenswrapper[4969]: I1004 09:46:07.073436 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" path="/var/lib/kubelet/pods/c0c99898-8c47-4801-855a-46f5e2f83669/volumes" Oct 04 09:46:10 crc kubenswrapper[4969]: I1004 09:46:10.056026 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:46:10 crc kubenswrapper[4969]: E1004 09:46:10.057000 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:46:24 crc kubenswrapper[4969]: I1004 09:46:24.056656 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:46:24 crc kubenswrapper[4969]: E1004 09:46:24.057889 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:46:35 crc kubenswrapper[4969]: I1004 09:46:35.055579 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:46:35 crc kubenswrapper[4969]: E1004 09:46:35.056990 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:46:47 crc kubenswrapper[4969]: I1004 09:46:47.056195 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:46:47 crc kubenswrapper[4969]: E1004 09:46:47.057320 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:47:01 crc kubenswrapper[4969]: I1004 09:47:01.055703 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:47:01 crc kubenswrapper[4969]: E1004 09:47:01.057345 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:47:12 crc kubenswrapper[4969]: I1004 09:47:12.055771 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:47:12 crc kubenswrapper[4969]: E1004 09:47:12.056638 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:47:23 crc kubenswrapper[4969]: I1004 09:47:23.063466 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:47:23 crc kubenswrapper[4969]: E1004 09:47:23.064239 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:47:36 crc kubenswrapper[4969]: I1004 09:47:36.055071 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:47:36 crc kubenswrapper[4969]: E1004 09:47:36.056248 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:47:49 crc kubenswrapper[4969]: I1004 09:47:49.055755 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:47:49 crc kubenswrapper[4969]: E1004 09:47:49.056989 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:48:00 crc kubenswrapper[4969]: I1004 09:48:00.055129 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:48:01 crc kubenswrapper[4969]: I1004 09:48:01.563840 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"85bb12a0449e14ed31dfed122d51d23546e416cdf9555e117a8d965fbe819adb"} Oct 04 09:50:19 crc kubenswrapper[4969]: I1004 09:50:19.666656 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:50:19 crc kubenswrapper[4969]: I1004 09:50:19.667352 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:50:49 crc kubenswrapper[4969]: I1004 09:50:49.666327 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:50:49 crc kubenswrapper[4969]: I1004 09:50:49.667114 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:51:19 crc kubenswrapper[4969]: I1004 09:51:19.666465 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:51:19 crc kubenswrapper[4969]: I1004 09:51:19.667188 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:51:19 crc kubenswrapper[4969]: I1004 09:51:19.667263 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:51:19 crc kubenswrapper[4969]: I1004 09:51:19.668133 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"85bb12a0449e14ed31dfed122d51d23546e416cdf9555e117a8d965fbe819adb"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:51:19 crc kubenswrapper[4969]: I1004 09:51:19.668234 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://85bb12a0449e14ed31dfed122d51d23546e416cdf9555e117a8d965fbe819adb" gracePeriod=600 Oct 04 09:51:19 crc kubenswrapper[4969]: I1004 09:51:19.959923 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="85bb12a0449e14ed31dfed122d51d23546e416cdf9555e117a8d965fbe819adb" exitCode=0 Oct 04 09:51:19 crc kubenswrapper[4969]: I1004 09:51:19.960069 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"85bb12a0449e14ed31dfed122d51d23546e416cdf9555e117a8d965fbe819adb"} Oct 04 09:51:19 crc kubenswrapper[4969]: I1004 09:51:19.960542 4969 scope.go:117] "RemoveContainer" containerID="7e9c48a5755dee3949f3f8ac0e9e2d67f8fe945082da6e5b1d5f4fedeee7e676" Oct 04 09:51:20 crc kubenswrapper[4969]: I1004 09:51:20.971179 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8"} Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.838911 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-znkhb"] Oct 04 09:51:48 crc kubenswrapper[4969]: E1004 09:51:48.840854 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" containerName="extract-content" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.840875 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" containerName="extract-content" Oct 04 09:51:48 crc kubenswrapper[4969]: E1004 09:51:48.840892 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerName="extract-utilities" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.840900 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerName="extract-utilities" Oct 04 09:51:48 crc kubenswrapper[4969]: E1004 09:51:48.840910 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerName="extract-content" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.840917 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerName="extract-content" Oct 04 09:51:48 crc kubenswrapper[4969]: E1004 09:51:48.840939 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" containerName="registry-server" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.840952 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" containerName="registry-server" Oct 04 09:51:48 crc kubenswrapper[4969]: E1004 09:51:48.840979 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerName="registry-server" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.840985 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerName="registry-server" Oct 04 09:51:48 crc kubenswrapper[4969]: E1004 09:51:48.841002 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" containerName="extract-utilities" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.841008 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" containerName="extract-utilities" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.841231 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0a4ae07-3f0f-47e9-ae40-8ecc79d08048" containerName="registry-server" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.841245 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0c99898-8c47-4801-855a-46f5e2f83669" containerName="registry-server" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.842651 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.857191 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-znkhb"] Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.994366 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-catalog-content\") pod \"certified-operators-znkhb\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.994701 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw6fr\" (UniqueName: \"kubernetes.io/projected/282ee3de-730c-4714-a750-f94a4b03711a-kube-api-access-dw6fr\") pod \"certified-operators-znkhb\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:48 crc kubenswrapper[4969]: I1004 09:51:48.994888 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-utilities\") pod \"certified-operators-znkhb\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:49 crc kubenswrapper[4969]: I1004 09:51:49.097090 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-catalog-content\") pod \"certified-operators-znkhb\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:49 crc kubenswrapper[4969]: I1004 09:51:49.097139 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw6fr\" (UniqueName: \"kubernetes.io/projected/282ee3de-730c-4714-a750-f94a4b03711a-kube-api-access-dw6fr\") pod \"certified-operators-znkhb\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:49 crc kubenswrapper[4969]: I1004 09:51:49.097192 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-utilities\") pod \"certified-operators-znkhb\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:49 crc kubenswrapper[4969]: I1004 09:51:49.097777 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-utilities\") pod \"certified-operators-znkhb\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:49 crc kubenswrapper[4969]: I1004 09:51:49.097986 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-catalog-content\") pod \"certified-operators-znkhb\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:49 crc kubenswrapper[4969]: I1004 09:51:49.119156 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw6fr\" (UniqueName: \"kubernetes.io/projected/282ee3de-730c-4714-a750-f94a4b03711a-kube-api-access-dw6fr\") pod \"certified-operators-znkhb\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:49 crc kubenswrapper[4969]: I1004 09:51:49.207233 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:49 crc kubenswrapper[4969]: I1004 09:51:49.773758 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-znkhb"] Oct 04 09:51:50 crc kubenswrapper[4969]: I1004 09:51:50.350591 4969 generic.go:334] "Generic (PLEG): container finished" podID="282ee3de-730c-4714-a750-f94a4b03711a" containerID="c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691" exitCode=0 Oct 04 09:51:50 crc kubenswrapper[4969]: I1004 09:51:50.350925 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znkhb" event={"ID":"282ee3de-730c-4714-a750-f94a4b03711a","Type":"ContainerDied","Data":"c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691"} Oct 04 09:51:50 crc kubenswrapper[4969]: I1004 09:51:50.350952 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znkhb" event={"ID":"282ee3de-730c-4714-a750-f94a4b03711a","Type":"ContainerStarted","Data":"d4ef9caaffddc9be659f7aefd6806c8f84c237c5c13c23be61fd9b7ebb1bdd5f"} Oct 04 09:51:50 crc kubenswrapper[4969]: I1004 09:51:50.354392 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 09:51:51 crc kubenswrapper[4969]: I1004 09:51:51.383126 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znkhb" event={"ID":"282ee3de-730c-4714-a750-f94a4b03711a","Type":"ContainerStarted","Data":"30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90"} Oct 04 09:51:53 crc kubenswrapper[4969]: I1004 09:51:53.408789 4969 generic.go:334] "Generic (PLEG): container finished" podID="282ee3de-730c-4714-a750-f94a4b03711a" containerID="30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90" exitCode=0 Oct 04 09:51:53 crc kubenswrapper[4969]: I1004 09:51:53.408941 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znkhb" event={"ID":"282ee3de-730c-4714-a750-f94a4b03711a","Type":"ContainerDied","Data":"30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90"} Oct 04 09:51:54 crc kubenswrapper[4969]: I1004 09:51:54.421944 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znkhb" event={"ID":"282ee3de-730c-4714-a750-f94a4b03711a","Type":"ContainerStarted","Data":"7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181"} Oct 04 09:51:54 crc kubenswrapper[4969]: I1004 09:51:54.458322 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-znkhb" podStartSLOduration=2.891336284 podStartE2EDuration="6.458298869s" podCreationTimestamp="2025-10-04 09:51:48 +0000 UTC" firstStartedPulling="2025-10-04 09:51:50.354207768 +0000 UTC m=+5738.108476582" lastFinishedPulling="2025-10-04 09:51:53.921170343 +0000 UTC m=+5741.675439167" observedRunningTime="2025-10-04 09:51:54.44095512 +0000 UTC m=+5742.195223964" watchObservedRunningTime="2025-10-04 09:51:54.458298869 +0000 UTC m=+5742.212567683" Oct 04 09:51:59 crc kubenswrapper[4969]: I1004 09:51:59.208059 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:59 crc kubenswrapper[4969]: I1004 09:51:59.208983 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:59 crc kubenswrapper[4969]: I1004 09:51:59.300111 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:59 crc kubenswrapper[4969]: I1004 09:51:59.537549 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:51:59 crc kubenswrapper[4969]: I1004 09:51:59.591574 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-znkhb"] Oct 04 09:52:01 crc kubenswrapper[4969]: I1004 09:52:01.519623 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-znkhb" podUID="282ee3de-730c-4714-a750-f94a4b03711a" containerName="registry-server" containerID="cri-o://7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181" gracePeriod=2 Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.071905 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.100804 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-utilities\") pod \"282ee3de-730c-4714-a750-f94a4b03711a\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.101007 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-catalog-content\") pod \"282ee3de-730c-4714-a750-f94a4b03711a\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.101054 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw6fr\" (UniqueName: \"kubernetes.io/projected/282ee3de-730c-4714-a750-f94a4b03711a-kube-api-access-dw6fr\") pod \"282ee3de-730c-4714-a750-f94a4b03711a\" (UID: \"282ee3de-730c-4714-a750-f94a4b03711a\") " Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.101841 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-utilities" (OuterVolumeSpecName: "utilities") pod "282ee3de-730c-4714-a750-f94a4b03711a" (UID: "282ee3de-730c-4714-a750-f94a4b03711a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.102227 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.115716 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/282ee3de-730c-4714-a750-f94a4b03711a-kube-api-access-dw6fr" (OuterVolumeSpecName: "kube-api-access-dw6fr") pod "282ee3de-730c-4714-a750-f94a4b03711a" (UID: "282ee3de-730c-4714-a750-f94a4b03711a"). InnerVolumeSpecName "kube-api-access-dw6fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.179314 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "282ee3de-730c-4714-a750-f94a4b03711a" (UID: "282ee3de-730c-4714-a750-f94a4b03711a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.205306 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282ee3de-730c-4714-a750-f94a4b03711a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.205358 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw6fr\" (UniqueName: \"kubernetes.io/projected/282ee3de-730c-4714-a750-f94a4b03711a-kube-api-access-dw6fr\") on node \"crc\" DevicePath \"\"" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.531482 4969 generic.go:334] "Generic (PLEG): container finished" podID="282ee3de-730c-4714-a750-f94a4b03711a" containerID="7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181" exitCode=0 Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.531680 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znkhb" event={"ID":"282ee3de-730c-4714-a750-f94a4b03711a","Type":"ContainerDied","Data":"7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181"} Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.531918 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znkhb" event={"ID":"282ee3de-730c-4714-a750-f94a4b03711a","Type":"ContainerDied","Data":"d4ef9caaffddc9be659f7aefd6806c8f84c237c5c13c23be61fd9b7ebb1bdd5f"} Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.531815 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znkhb" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.532029 4969 scope.go:117] "RemoveContainer" containerID="7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.566246 4969 scope.go:117] "RemoveContainer" containerID="30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.584362 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-znkhb"] Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.597890 4969 scope.go:117] "RemoveContainer" containerID="c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.598343 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-znkhb"] Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.675157 4969 scope.go:117] "RemoveContainer" containerID="7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181" Oct 04 09:52:02 crc kubenswrapper[4969]: E1004 09:52:02.675557 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181\": container with ID starting with 7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181 not found: ID does not exist" containerID="7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.675593 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181"} err="failed to get container status \"7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181\": rpc error: code = NotFound desc = could not find container \"7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181\": container with ID starting with 7060261a24d7bbcf0ad8be4ac24f84f98495f579a03eeedb72373cbb0ec37181 not found: ID does not exist" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.675617 4969 scope.go:117] "RemoveContainer" containerID="30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90" Oct 04 09:52:02 crc kubenswrapper[4969]: E1004 09:52:02.676184 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90\": container with ID starting with 30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90 not found: ID does not exist" containerID="30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.676240 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90"} err="failed to get container status \"30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90\": rpc error: code = NotFound desc = could not find container \"30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90\": container with ID starting with 30dab2a5ab37d76ce03a6812bc4a9bf40a641d0117e71cc1860ce4c56c384e90 not found: ID does not exist" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.676276 4969 scope.go:117] "RemoveContainer" containerID="c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691" Oct 04 09:52:02 crc kubenswrapper[4969]: E1004 09:52:02.677070 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691\": container with ID starting with c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691 not found: ID does not exist" containerID="c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691" Oct 04 09:52:02 crc kubenswrapper[4969]: I1004 09:52:02.677102 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691"} err="failed to get container status \"c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691\": rpc error: code = NotFound desc = could not find container \"c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691\": container with ID starting with c917115326b4fa1e65cc7c1bdaaef4ee763a034aeaa005174b8c5561e7a76691 not found: ID does not exist" Oct 04 09:52:03 crc kubenswrapper[4969]: I1004 09:52:03.071975 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="282ee3de-730c-4714-a750-f94a4b03711a" path="/var/lib/kubelet/pods/282ee3de-730c-4714-a750-f94a4b03711a/volumes" Oct 04 09:53:49 crc kubenswrapper[4969]: I1004 09:53:49.666153 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:53:49 crc kubenswrapper[4969]: I1004 09:53:49.666831 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:53:51 crc kubenswrapper[4969]: I1004 09:53:51.929832 4969 generic.go:334] "Generic (PLEG): container finished" podID="6348f782-6e3b-4c67-b252-2276628c14d6" containerID="95da218138474e9dafb1c5e2f0055868b9b87f05b47cec19c4abea56f80ae0ed" exitCode=0 Oct 04 09:53:51 crc kubenswrapper[4969]: I1004 09:53:51.930652 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"6348f782-6e3b-4c67-b252-2276628c14d6","Type":"ContainerDied","Data":"95da218138474e9dafb1c5e2f0055868b9b87f05b47cec19c4abea56f80ae0ed"} Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.380209 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.439460 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config-secret\") pod \"6348f782-6e3b-4c67-b252-2276628c14d6\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.439510 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-workdir\") pod \"6348f782-6e3b-4c67-b252-2276628c14d6\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.439544 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ssh-key\") pod \"6348f782-6e3b-4c67-b252-2276628c14d6\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.439573 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-temporary\") pod \"6348f782-6e3b-4c67-b252-2276628c14d6\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.439622 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"6348f782-6e3b-4c67-b252-2276628c14d6\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.439647 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config\") pod \"6348f782-6e3b-4c67-b252-2276628c14d6\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.439712 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-config-data\") pod \"6348f782-6e3b-4c67-b252-2276628c14d6\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.439803 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqw2t\" (UniqueName: \"kubernetes.io/projected/6348f782-6e3b-4c67-b252-2276628c14d6-kube-api-access-qqw2t\") pod \"6348f782-6e3b-4c67-b252-2276628c14d6\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.439832 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ca-certs\") pod \"6348f782-6e3b-4c67-b252-2276628c14d6\" (UID: \"6348f782-6e3b-4c67-b252-2276628c14d6\") " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.440367 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "6348f782-6e3b-4c67-b252-2276628c14d6" (UID: "6348f782-6e3b-4c67-b252-2276628c14d6"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.440821 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-config-data" (OuterVolumeSpecName: "config-data") pod "6348f782-6e3b-4c67-b252-2276628c14d6" (UID: "6348f782-6e3b-4c67-b252-2276628c14d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.443901 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "6348f782-6e3b-4c67-b252-2276628c14d6" (UID: "6348f782-6e3b-4c67-b252-2276628c14d6"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.459305 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "6348f782-6e3b-4c67-b252-2276628c14d6" (UID: "6348f782-6e3b-4c67-b252-2276628c14d6"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.461379 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6348f782-6e3b-4c67-b252-2276628c14d6-kube-api-access-qqw2t" (OuterVolumeSpecName: "kube-api-access-qqw2t") pod "6348f782-6e3b-4c67-b252-2276628c14d6" (UID: "6348f782-6e3b-4c67-b252-2276628c14d6"). InnerVolumeSpecName "kube-api-access-qqw2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.472572 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6348f782-6e3b-4c67-b252-2276628c14d6" (UID: "6348f782-6e3b-4c67-b252-2276628c14d6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.475560 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "6348f782-6e3b-4c67-b252-2276628c14d6" (UID: "6348f782-6e3b-4c67-b252-2276628c14d6"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.503504 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "6348f782-6e3b-4c67-b252-2276628c14d6" (UID: "6348f782-6e3b-4c67-b252-2276628c14d6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.514677 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "6348f782-6e3b-4c67-b252-2276628c14d6" (UID: "6348f782-6e3b-4c67-b252-2276628c14d6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.541944 4969 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.542013 4969 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.542028 4969 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.542044 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6348f782-6e3b-4c67-b252-2276628c14d6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.542056 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqw2t\" (UniqueName: \"kubernetes.io/projected/6348f782-6e3b-4c67-b252-2276628c14d6-kube-api-access-qqw2t\") on node \"crc\" DevicePath \"\"" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.542066 4969 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.542077 4969 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.542088 4969 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6348f782-6e3b-4c67-b252-2276628c14d6-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.542098 4969 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6348f782-6e3b-4c67-b252-2276628c14d6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.578795 4969 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.644620 4969 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.958076 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"6348f782-6e3b-4c67-b252-2276628c14d6","Type":"ContainerDied","Data":"ac15602ac98165544941c206deea5027ba468ded13f8a6c01ffc5b7dae08684a"} Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.958138 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac15602ac98165544941c206deea5027ba468ded13f8a6c01ffc5b7dae08684a" Oct 04 09:53:53 crc kubenswrapper[4969]: I1004 09:53:53.958249 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.121194 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 09:53:56 crc kubenswrapper[4969]: E1004 09:53:56.126615 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282ee3de-730c-4714-a750-f94a4b03711a" containerName="registry-server" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.126668 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="282ee3de-730c-4714-a750-f94a4b03711a" containerName="registry-server" Oct 04 09:53:56 crc kubenswrapper[4969]: E1004 09:53:56.126708 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282ee3de-730c-4714-a750-f94a4b03711a" containerName="extract-utilities" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.126724 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="282ee3de-730c-4714-a750-f94a4b03711a" containerName="extract-utilities" Oct 04 09:53:56 crc kubenswrapper[4969]: E1004 09:53:56.126753 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6348f782-6e3b-4c67-b252-2276628c14d6" containerName="tempest-tests-tempest-tests-runner" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.126767 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6348f782-6e3b-4c67-b252-2276628c14d6" containerName="tempest-tests-tempest-tests-runner" Oct 04 09:53:56 crc kubenswrapper[4969]: E1004 09:53:56.126798 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282ee3de-730c-4714-a750-f94a4b03711a" containerName="extract-content" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.126812 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="282ee3de-730c-4714-a750-f94a4b03711a" containerName="extract-content" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.127634 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="282ee3de-730c-4714-a750-f94a4b03711a" containerName="registry-server" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.127671 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="6348f782-6e3b-4c67-b252-2276628c14d6" containerName="tempest-tests-tempest-tests-runner" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.128801 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.131773 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gh4r4" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.145556 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.313694 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj8ct\" (UniqueName: \"kubernetes.io/projected/9e320e23-ff99-4d9f-b2aa-25c7d2420b88-kube-api-access-vj8ct\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9e320e23-ff99-4d9f-b2aa-25c7d2420b88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.313803 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9e320e23-ff99-4d9f-b2aa-25c7d2420b88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.416561 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj8ct\" (UniqueName: \"kubernetes.io/projected/9e320e23-ff99-4d9f-b2aa-25c7d2420b88-kube-api-access-vj8ct\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9e320e23-ff99-4d9f-b2aa-25c7d2420b88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.417015 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9e320e23-ff99-4d9f-b2aa-25c7d2420b88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.417708 4969 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9e320e23-ff99-4d9f-b2aa-25c7d2420b88\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.440290 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj8ct\" (UniqueName: \"kubernetes.io/projected/9e320e23-ff99-4d9f-b2aa-25c7d2420b88-kube-api-access-vj8ct\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9e320e23-ff99-4d9f-b2aa-25c7d2420b88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.452688 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9e320e23-ff99-4d9f-b2aa-25c7d2420b88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 09:53:56 crc kubenswrapper[4969]: I1004 09:53:56.458930 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 09:53:57 crc kubenswrapper[4969]: I1004 09:53:57.047918 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 09:53:58 crc kubenswrapper[4969]: I1004 09:53:58.013144 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"9e320e23-ff99-4d9f-b2aa-25c7d2420b88","Type":"ContainerStarted","Data":"5b3d8d8ab026b1fbe34aad86f1ff6a88482ff4fdd4dd2abc1ccbf63002a85b78"} Oct 04 09:53:59 crc kubenswrapper[4969]: I1004 09:53:59.037444 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"9e320e23-ff99-4d9f-b2aa-25c7d2420b88","Type":"ContainerStarted","Data":"def1a0592c4bbece038ba80fc507bc26d8233ffb50bc8f54c89b6bb8d97c527e"} Oct 04 09:53:59 crc kubenswrapper[4969]: I1004 09:53:59.051877 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.989918287 podStartE2EDuration="3.051858775s" podCreationTimestamp="2025-10-04 09:53:56 +0000 UTC" firstStartedPulling="2025-10-04 09:53:57.056166103 +0000 UTC m=+5864.810434917" lastFinishedPulling="2025-10-04 09:53:58.118106581 +0000 UTC m=+5865.872375405" observedRunningTime="2025-10-04 09:53:59.051756933 +0000 UTC m=+5866.806025757" watchObservedRunningTime="2025-10-04 09:53:59.051858775 +0000 UTC m=+5866.806127589" Oct 04 09:54:15 crc kubenswrapper[4969]: I1004 09:54:15.938040 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mtfk4/must-gather-xz6qb"] Oct 04 09:54:15 crc kubenswrapper[4969]: I1004 09:54:15.940343 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 09:54:15 crc kubenswrapper[4969]: I1004 09:54:15.947411 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mtfk4/must-gather-xz6qb"] Oct 04 09:54:15 crc kubenswrapper[4969]: I1004 09:54:15.947583 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mtfk4"/"openshift-service-ca.crt" Oct 04 09:54:15 crc kubenswrapper[4969]: I1004 09:54:15.948105 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mtfk4"/"kube-root-ca.crt" Oct 04 09:54:15 crc kubenswrapper[4969]: I1004 09:54:15.948613 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mtfk4"/"default-dockercfg-96c5m" Oct 04 09:54:16 crc kubenswrapper[4969]: I1004 09:54:16.068517 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj9k5\" (UniqueName: \"kubernetes.io/projected/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-kube-api-access-cj9k5\") pod \"must-gather-xz6qb\" (UID: \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\") " pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 09:54:16 crc kubenswrapper[4969]: I1004 09:54:16.068595 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-must-gather-output\") pod \"must-gather-xz6qb\" (UID: \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\") " pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 09:54:16 crc kubenswrapper[4969]: I1004 09:54:16.170470 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj9k5\" (UniqueName: \"kubernetes.io/projected/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-kube-api-access-cj9k5\") pod \"must-gather-xz6qb\" (UID: \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\") " pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 09:54:16 crc kubenswrapper[4969]: I1004 09:54:16.170587 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-must-gather-output\") pod \"must-gather-xz6qb\" (UID: \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\") " pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 09:54:16 crc kubenswrapper[4969]: I1004 09:54:16.171077 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-must-gather-output\") pod \"must-gather-xz6qb\" (UID: \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\") " pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 09:54:16 crc kubenswrapper[4969]: I1004 09:54:16.203467 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj9k5\" (UniqueName: \"kubernetes.io/projected/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-kube-api-access-cj9k5\") pod \"must-gather-xz6qb\" (UID: \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\") " pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 09:54:16 crc kubenswrapper[4969]: I1004 09:54:16.257312 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 09:54:16 crc kubenswrapper[4969]: I1004 09:54:16.849837 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mtfk4/must-gather-xz6qb"] Oct 04 09:54:17 crc kubenswrapper[4969]: I1004 09:54:17.241191 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" event={"ID":"042bd31c-d47c-488f-b1e6-4765f3b3cbc8","Type":"ContainerStarted","Data":"0933ee7780b9a4203622ac783b2f3cb6d8f64004486879a3e78c37202b6faf85"} Oct 04 09:54:19 crc kubenswrapper[4969]: I1004 09:54:19.666209 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:54:19 crc kubenswrapper[4969]: I1004 09:54:19.666639 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:54:24 crc kubenswrapper[4969]: I1004 09:54:24.317003 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" event={"ID":"042bd31c-d47c-488f-b1e6-4765f3b3cbc8","Type":"ContainerStarted","Data":"13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b"} Oct 04 09:54:24 crc kubenswrapper[4969]: I1004 09:54:24.317714 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" event={"ID":"042bd31c-d47c-488f-b1e6-4765f3b3cbc8","Type":"ContainerStarted","Data":"0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905"} Oct 04 09:54:24 crc kubenswrapper[4969]: I1004 09:54:24.389616 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" podStartSLOduration=2.925557746 podStartE2EDuration="9.389583158s" podCreationTimestamp="2025-10-04 09:54:15 +0000 UTC" firstStartedPulling="2025-10-04 09:54:16.851707352 +0000 UTC m=+5884.605976166" lastFinishedPulling="2025-10-04 09:54:23.315732764 +0000 UTC m=+5891.070001578" observedRunningTime="2025-10-04 09:54:24.376196822 +0000 UTC m=+5892.130465636" watchObservedRunningTime="2025-10-04 09:54:24.389583158 +0000 UTC m=+5892.143852022" Oct 04 09:54:27 crc kubenswrapper[4969]: I1004 09:54:27.191939 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mtfk4/crc-debug-xprp7"] Oct 04 09:54:27 crc kubenswrapper[4969]: I1004 09:54:27.197973 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:54:27 crc kubenswrapper[4969]: I1004 09:54:27.302056 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbqrw\" (UniqueName: \"kubernetes.io/projected/3588eff0-a2f0-4233-9f16-a537bef351c3-kube-api-access-pbqrw\") pod \"crc-debug-xprp7\" (UID: \"3588eff0-a2f0-4233-9f16-a537bef351c3\") " pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:54:27 crc kubenswrapper[4969]: I1004 09:54:27.302461 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3588eff0-a2f0-4233-9f16-a537bef351c3-host\") pod \"crc-debug-xprp7\" (UID: \"3588eff0-a2f0-4233-9f16-a537bef351c3\") " pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:54:27 crc kubenswrapper[4969]: I1004 09:54:27.404561 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3588eff0-a2f0-4233-9f16-a537bef351c3-host\") pod \"crc-debug-xprp7\" (UID: \"3588eff0-a2f0-4233-9f16-a537bef351c3\") " pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:54:27 crc kubenswrapper[4969]: I1004 09:54:27.404698 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3588eff0-a2f0-4233-9f16-a537bef351c3-host\") pod \"crc-debug-xprp7\" (UID: \"3588eff0-a2f0-4233-9f16-a537bef351c3\") " pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:54:27 crc kubenswrapper[4969]: I1004 09:54:27.405035 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbqrw\" (UniqueName: \"kubernetes.io/projected/3588eff0-a2f0-4233-9f16-a537bef351c3-kube-api-access-pbqrw\") pod \"crc-debug-xprp7\" (UID: \"3588eff0-a2f0-4233-9f16-a537bef351c3\") " pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:54:27 crc kubenswrapper[4969]: I1004 09:54:27.423193 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbqrw\" (UniqueName: \"kubernetes.io/projected/3588eff0-a2f0-4233-9f16-a537bef351c3-kube-api-access-pbqrw\") pod \"crc-debug-xprp7\" (UID: \"3588eff0-a2f0-4233-9f16-a537bef351c3\") " pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:54:27 crc kubenswrapper[4969]: I1004 09:54:27.515772 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:54:28 crc kubenswrapper[4969]: I1004 09:54:28.350847 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/crc-debug-xprp7" event={"ID":"3588eff0-a2f0-4233-9f16-a537bef351c3","Type":"ContainerStarted","Data":"b6882ef7a47ed2d21d144ca1ca8567c08cb1bd0c794b5d10b916ba24eff20927"} Oct 04 09:54:38 crc kubenswrapper[4969]: I1004 09:54:38.445199 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/crc-debug-xprp7" event={"ID":"3588eff0-a2f0-4233-9f16-a537bef351c3","Type":"ContainerStarted","Data":"f1110b04e100376a6109608c535768757d1c9e86735425c3f6e92dd519265321"} Oct 04 09:54:38 crc kubenswrapper[4969]: I1004 09:54:38.489687 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mtfk4/crc-debug-xprp7" podStartSLOduration=1.282039905 podStartE2EDuration="11.48966904s" podCreationTimestamp="2025-10-04 09:54:27 +0000 UTC" firstStartedPulling="2025-10-04 09:54:27.589573147 +0000 UTC m=+5895.343841961" lastFinishedPulling="2025-10-04 09:54:37.797202272 +0000 UTC m=+5905.551471096" observedRunningTime="2025-10-04 09:54:38.475202981 +0000 UTC m=+5906.229471795" watchObservedRunningTime="2025-10-04 09:54:38.48966904 +0000 UTC m=+5906.243937844" Oct 04 09:54:49 crc kubenswrapper[4969]: I1004 09:54:49.667776 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 09:54:49 crc kubenswrapper[4969]: I1004 09:54:49.668546 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 09:54:49 crc kubenswrapper[4969]: I1004 09:54:49.668607 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 09:54:49 crc kubenswrapper[4969]: I1004 09:54:49.669511 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 09:54:49 crc kubenswrapper[4969]: I1004 09:54:49.669603 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" gracePeriod=600 Oct 04 09:54:50 crc kubenswrapper[4969]: E1004 09:54:50.091855 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:54:50 crc kubenswrapper[4969]: I1004 09:54:50.590450 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" exitCode=0 Oct 04 09:54:50 crc kubenswrapper[4969]: I1004 09:54:50.590646 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8"} Oct 04 09:54:50 crc kubenswrapper[4969]: I1004 09:54:50.590969 4969 scope.go:117] "RemoveContainer" containerID="85bb12a0449e14ed31dfed122d51d23546e416cdf9555e117a8d965fbe819adb" Oct 04 09:54:50 crc kubenswrapper[4969]: I1004 09:54:50.591866 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:54:50 crc kubenswrapper[4969]: E1004 09:54:50.592342 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:55:05 crc kubenswrapper[4969]: I1004 09:55:05.055513 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:55:05 crc kubenswrapper[4969]: E1004 09:55:05.056411 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:55:15 crc kubenswrapper[4969]: I1004 09:55:15.862855 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fcfm5"] Oct 04 09:55:15 crc kubenswrapper[4969]: I1004 09:55:15.867239 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:15 crc kubenswrapper[4969]: I1004 09:55:15.877399 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fcfm5"] Oct 04 09:55:15 crc kubenswrapper[4969]: I1004 09:55:15.992188 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-catalog-content\") pod \"community-operators-fcfm5\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:15 crc kubenswrapper[4969]: I1004 09:55:15.993815 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-utilities\") pod \"community-operators-fcfm5\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:15 crc kubenswrapper[4969]: I1004 09:55:15.993863 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ns9h\" (UniqueName: \"kubernetes.io/projected/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-kube-api-access-2ns9h\") pod \"community-operators-fcfm5\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:16 crc kubenswrapper[4969]: I1004 09:55:16.095580 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-catalog-content\") pod \"community-operators-fcfm5\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:16 crc kubenswrapper[4969]: I1004 09:55:16.095881 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-utilities\") pod \"community-operators-fcfm5\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:16 crc kubenswrapper[4969]: I1004 09:55:16.095903 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ns9h\" (UniqueName: \"kubernetes.io/projected/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-kube-api-access-2ns9h\") pod \"community-operators-fcfm5\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:16 crc kubenswrapper[4969]: I1004 09:55:16.096470 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-utilities\") pod \"community-operators-fcfm5\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:16 crc kubenswrapper[4969]: I1004 09:55:16.096470 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-catalog-content\") pod \"community-operators-fcfm5\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:16 crc kubenswrapper[4969]: I1004 09:55:16.124195 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ns9h\" (UniqueName: \"kubernetes.io/projected/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-kube-api-access-2ns9h\") pod \"community-operators-fcfm5\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:16 crc kubenswrapper[4969]: I1004 09:55:16.237656 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:16 crc kubenswrapper[4969]: I1004 09:55:16.825933 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fcfm5"] Oct 04 09:55:16 crc kubenswrapper[4969]: I1004 09:55:16.892780 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcfm5" event={"ID":"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d","Type":"ContainerStarted","Data":"fb84f8614446ba76125f5df1210f3dd5e1984f6986363e4daa77a77fc8d6ebe1"} Oct 04 09:55:17 crc kubenswrapper[4969]: I1004 09:55:17.903460 4969 generic.go:334] "Generic (PLEG): container finished" podID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerID="242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670" exitCode=0 Oct 04 09:55:17 crc kubenswrapper[4969]: I1004 09:55:17.903557 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcfm5" event={"ID":"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d","Type":"ContainerDied","Data":"242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670"} Oct 04 09:55:18 crc kubenswrapper[4969]: I1004 09:55:18.920962 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcfm5" event={"ID":"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d","Type":"ContainerStarted","Data":"94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272"} Oct 04 09:55:20 crc kubenswrapper[4969]: I1004 09:55:20.055259 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:55:20 crc kubenswrapper[4969]: E1004 09:55:20.055925 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:55:20 crc kubenswrapper[4969]: I1004 09:55:20.938645 4969 generic.go:334] "Generic (PLEG): container finished" podID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerID="94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272" exitCode=0 Oct 04 09:55:20 crc kubenswrapper[4969]: I1004 09:55:20.938690 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcfm5" event={"ID":"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d","Type":"ContainerDied","Data":"94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272"} Oct 04 09:55:22 crc kubenswrapper[4969]: I1004 09:55:22.954686 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcfm5" event={"ID":"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d","Type":"ContainerStarted","Data":"9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a"} Oct 04 09:55:22 crc kubenswrapper[4969]: I1004 09:55:22.971314 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fcfm5" podStartSLOduration=3.5101144 podStartE2EDuration="7.971297772s" podCreationTimestamp="2025-10-04 09:55:15 +0000 UTC" firstStartedPulling="2025-10-04 09:55:17.906041581 +0000 UTC m=+5945.660310405" lastFinishedPulling="2025-10-04 09:55:22.367224963 +0000 UTC m=+5950.121493777" observedRunningTime="2025-10-04 09:55:22.970998365 +0000 UTC m=+5950.725267179" watchObservedRunningTime="2025-10-04 09:55:22.971297772 +0000 UTC m=+5950.725566586" Oct 04 09:55:26 crc kubenswrapper[4969]: I1004 09:55:26.237882 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:26 crc kubenswrapper[4969]: I1004 09:55:26.238375 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:27 crc kubenswrapper[4969]: I1004 09:55:27.285888 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-fcfm5" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerName="registry-server" probeResult="failure" output=< Oct 04 09:55:27 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:55:27 crc kubenswrapper[4969]: > Oct 04 09:55:31 crc kubenswrapper[4969]: I1004 09:55:31.055125 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:55:31 crc kubenswrapper[4969]: E1004 09:55:31.055815 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:55:36 crc kubenswrapper[4969]: I1004 09:55:36.291983 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:36 crc kubenswrapper[4969]: I1004 09:55:36.355531 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:36 crc kubenswrapper[4969]: I1004 09:55:36.529792 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fcfm5"] Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.109742 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fcfm5" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerName="registry-server" containerID="cri-o://9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a" gracePeriod=2 Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.624967 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.798723 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ns9h\" (UniqueName: \"kubernetes.io/projected/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-kube-api-access-2ns9h\") pod \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.798856 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-utilities\") pod \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.798947 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-catalog-content\") pod \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\" (UID: \"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d\") " Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.799725 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-utilities" (OuterVolumeSpecName: "utilities") pod "6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" (UID: "6396e6f8-d2a1-45bd-aae3-cef656c0ef0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.800059 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.806123 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-kube-api-access-2ns9h" (OuterVolumeSpecName: "kube-api-access-2ns9h") pod "6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" (UID: "6396e6f8-d2a1-45bd-aae3-cef656c0ef0d"). InnerVolumeSpecName "kube-api-access-2ns9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.853678 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" (UID: "6396e6f8-d2a1-45bd-aae3-cef656c0ef0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.902651 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ns9h\" (UniqueName: \"kubernetes.io/projected/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-kube-api-access-2ns9h\") on node \"crc\" DevicePath \"\"" Oct 04 09:55:38 crc kubenswrapper[4969]: I1004 09:55:38.902690 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.124677 4969 generic.go:334] "Generic (PLEG): container finished" podID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerID="9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a" exitCode=0 Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.124723 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcfm5" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.124763 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcfm5" event={"ID":"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d","Type":"ContainerDied","Data":"9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a"} Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.125151 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcfm5" event={"ID":"6396e6f8-d2a1-45bd-aae3-cef656c0ef0d","Type":"ContainerDied","Data":"fb84f8614446ba76125f5df1210f3dd5e1984f6986363e4daa77a77fc8d6ebe1"} Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.125174 4969 scope.go:117] "RemoveContainer" containerID="9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.158677 4969 scope.go:117] "RemoveContainer" containerID="94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.164670 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fcfm5"] Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.179645 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fcfm5"] Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.189150 4969 scope.go:117] "RemoveContainer" containerID="242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.241768 4969 scope.go:117] "RemoveContainer" containerID="9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a" Oct 04 09:55:39 crc kubenswrapper[4969]: E1004 09:55:39.242194 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a\": container with ID starting with 9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a not found: ID does not exist" containerID="9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.242305 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a"} err="failed to get container status \"9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a\": rpc error: code = NotFound desc = could not find container \"9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a\": container with ID starting with 9d60882146643cfdc53bd782d0247c30bc01d953522da8c2f2ba5a410699276a not found: ID does not exist" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.242339 4969 scope.go:117] "RemoveContainer" containerID="94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272" Oct 04 09:55:39 crc kubenswrapper[4969]: E1004 09:55:39.242813 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272\": container with ID starting with 94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272 not found: ID does not exist" containerID="94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.242862 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272"} err="failed to get container status \"94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272\": rpc error: code = NotFound desc = could not find container \"94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272\": container with ID starting with 94e07e35dd13b084996ed8be627453f6fbdb37522781d80c6a43f4a6c1ceb272 not found: ID does not exist" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.242888 4969 scope.go:117] "RemoveContainer" containerID="242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670" Oct 04 09:55:39 crc kubenswrapper[4969]: E1004 09:55:39.243133 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670\": container with ID starting with 242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670 not found: ID does not exist" containerID="242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670" Oct 04 09:55:39 crc kubenswrapper[4969]: I1004 09:55:39.243165 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670"} err="failed to get container status \"242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670\": rpc error: code = NotFound desc = could not find container \"242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670\": container with ID starting with 242b2b00258b65287c2dbcafdb90b7d69176da6a003d5769e4e92dc53ca42670 not found: ID does not exist" Oct 04 09:55:41 crc kubenswrapper[4969]: I1004 09:55:41.067879 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" path="/var/lib/kubelet/pods/6396e6f8-d2a1-45bd-aae3-cef656c0ef0d/volumes" Oct 04 09:55:43 crc kubenswrapper[4969]: I1004 09:55:43.064113 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:55:43 crc kubenswrapper[4969]: E1004 09:55:43.065062 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:55:49 crc kubenswrapper[4969]: I1004 09:55:49.107091 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dbff44956-qlvk2_d46c925d-1b2f-4696-b099-f3a549dda18b/barbican-api/0.log" Oct 04 09:55:49 crc kubenswrapper[4969]: I1004 09:55:49.123446 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dbff44956-qlvk2_d46c925d-1b2f-4696-b099-f3a549dda18b/barbican-api-log/0.log" Oct 04 09:55:49 crc kubenswrapper[4969]: I1004 09:55:49.347166 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7d5b5f7c58-kgs6c_f2e7b743-71f9-46c8-aeb7-450c63bf22ca/barbican-keystone-listener/0.log" Oct 04 09:55:49 crc kubenswrapper[4969]: I1004 09:55:49.379638 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7d5b5f7c58-kgs6c_f2e7b743-71f9-46c8-aeb7-450c63bf22ca/barbican-keystone-listener-log/0.log" Oct 04 09:55:49 crc kubenswrapper[4969]: I1004 09:55:49.536046 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9fb8f975c-4l4lz_3f74b083-ba83-4f80-a8a6-e5e10e6dde28/barbican-worker/0.log" Oct 04 09:55:49 crc kubenswrapper[4969]: I1004 09:55:49.564958 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9fb8f975c-4l4lz_3f74b083-ba83-4f80-a8a6-e5e10e6dde28/barbican-worker-log/0.log" Oct 04 09:55:49 crc kubenswrapper[4969]: I1004 09:55:49.769395 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw_6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:50 crc kubenswrapper[4969]: I1004 09:55:50.007507 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/ceilometer-central-agent/1.log" Oct 04 09:55:50 crc kubenswrapper[4969]: I1004 09:55:50.012267 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/ceilometer-central-agent/0.log" Oct 04 09:55:50 crc kubenswrapper[4969]: I1004 09:55:50.089364 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/ceilometer-notification-agent/0.log" Oct 04 09:55:50 crc kubenswrapper[4969]: I1004 09:55:50.201126 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/sg-core/0.log" Oct 04 09:55:50 crc kubenswrapper[4969]: I1004 09:55:50.216493 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/proxy-httpd/0.log" Oct 04 09:55:50 crc kubenswrapper[4969]: I1004 09:55:50.444827 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_755c8160-b3df-4ba0-8d91-46d4af514dfc/cinder-api-log/0.log" Oct 04 09:55:50 crc kubenswrapper[4969]: I1004 09:55:50.617673 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_755c8160-b3df-4ba0-8d91-46d4af514dfc/cinder-api/0.log" Oct 04 09:55:50 crc kubenswrapper[4969]: I1004 09:55:50.713973 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a7183f88-c473-4e74-8010-f8e7a380b35f/cinder-scheduler/0.log" Oct 04 09:55:50 crc kubenswrapper[4969]: I1004 09:55:50.837577 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a7183f88-c473-4e74-8010-f8e7a380b35f/probe/0.log" Oct 04 09:55:51 crc kubenswrapper[4969]: I1004 09:55:51.000192 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2_48e0d7b9-c122-4cee-84b3-1e89a72af2ad/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:51 crc kubenswrapper[4969]: I1004 09:55:51.167759 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk_85fdbfa3-25a5-4a92-9473-198d3cf416c8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:51 crc kubenswrapper[4969]: I1004 09:55:51.346649 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-t796n_19b4ae96-fa76-4d96-b4e0-d0058345fc23/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:51 crc kubenswrapper[4969]: I1004 09:55:51.466541 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bd8b95f47-sjpqk_342bf40f-63d2-47cd-82c8-5c2aec79d3d2/init/0.log" Oct 04 09:55:51 crc kubenswrapper[4969]: I1004 09:55:51.696997 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bd8b95f47-sjpqk_342bf40f-63d2-47cd-82c8-5c2aec79d3d2/init/0.log" Oct 04 09:55:51 crc kubenswrapper[4969]: I1004 09:55:51.879686 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bd8b95f47-sjpqk_342bf40f-63d2-47cd-82c8-5c2aec79d3d2/dnsmasq-dns/0.log" Oct 04 09:55:51 crc kubenswrapper[4969]: I1004 09:55:51.927702 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz_a71c86d5-acae-401d-bea7-8cee0956d990/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:52 crc kubenswrapper[4969]: I1004 09:55:52.088384 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7c627df9-6989-44d8-b72f-128992d65bd9/glance-httpd/0.log" Oct 04 09:55:52 crc kubenswrapper[4969]: I1004 09:55:52.126351 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7c627df9-6989-44d8-b72f-128992d65bd9/glance-log/0.log" Oct 04 09:55:52 crc kubenswrapper[4969]: I1004 09:55:52.278646 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7ebb3f7a-584d-4c2d-a26f-f66481883b6e/glance-httpd/0.log" Oct 04 09:55:52 crc kubenswrapper[4969]: I1004 09:55:52.315239 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7ebb3f7a-584d-4c2d-a26f-f66481883b6e/glance-log/0.log" Oct 04 09:55:52 crc kubenswrapper[4969]: I1004 09:55:52.504886 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f94787ff8-mkj4z_3344b512-4dea-4112-a8b1-7fa337ef81ae/horizon/0.log" Oct 04 09:55:52 crc kubenswrapper[4969]: I1004 09:55:52.659418 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw_a7f1a5ed-9324-4eb1-b561-d65fca042f4d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:52 crc kubenswrapper[4969]: I1004 09:55:52.776010 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-cqqr6_579fed39-17e4-412c-b76b-139db1b5938e/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:53 crc kubenswrapper[4969]: I1004 09:55:53.094659 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f94787ff8-mkj4z_3344b512-4dea-4112-a8b1-7fa337ef81ae/horizon-log/0.log" Oct 04 09:55:53 crc kubenswrapper[4969]: I1004 09:55:53.175644 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29326141-knk55_cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a/keystone-cron/0.log" Oct 04 09:55:53 crc kubenswrapper[4969]: I1004 09:55:53.381766 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_5385fdd1-eb06-4d69-a383-236a55505108/kube-state-metrics/0.log" Oct 04 09:55:53 crc kubenswrapper[4969]: I1004 09:55:53.392056 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-699ff78d6b-dv46q_7327643b-aa39-4085-88c4-bc4411b8d832/keystone-api/0.log" Oct 04 09:55:53 crc kubenswrapper[4969]: I1004 09:55:53.552987 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9_07ace2b0-20d8-4b9a-8f93-4649a3eee5bb/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:54 crc kubenswrapper[4969]: I1004 09:55:54.114751 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-78c8455845-gfrwq_e82aaa27-6e5c-43d8-9ed2-c0958de485c7/neutron-httpd/0.log" Oct 04 09:55:54 crc kubenswrapper[4969]: I1004 09:55:54.143729 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-78c8455845-gfrwq_e82aaa27-6e5c-43d8-9ed2-c0958de485c7/neutron-api/0.log" Oct 04 09:55:54 crc kubenswrapper[4969]: I1004 09:55:54.196283 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4_87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:55 crc kubenswrapper[4969]: I1004 09:55:55.087616 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_60f12325-d22b-4986-a653-854d290f0adc/nova-cell0-conductor-conductor/0.log" Oct 04 09:55:55 crc kubenswrapper[4969]: I1004 09:55:55.646345 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ac48f46b-0304-424d-b394-a1e2a274230a/nova-cell1-conductor-conductor/0.log" Oct 04 09:55:55 crc kubenswrapper[4969]: I1004 09:55:55.909312 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e5b33466-d882-4113-8c9a-bc1047ee6b9a/nova-api-log/0.log" Oct 04 09:55:56 crc kubenswrapper[4969]: I1004 09:55:56.143946 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_41d6e66e-6079-4278-a26c-04a299dd6d9d/memcached/0.log" Oct 04 09:55:56 crc kubenswrapper[4969]: I1004 09:55:56.160225 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e5b33466-d882-4113-8c9a-bc1047ee6b9a/nova-api-api/0.log" Oct 04 09:55:56 crc kubenswrapper[4969]: I1004 09:55:56.164498 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_2075dd48-ec13-4ebb-841b-3a2d82724f98/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 09:55:56 crc kubenswrapper[4969]: I1004 09:55:56.445685 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-p9v4f_4ce3e260-2be8-435b-8a24-7d06614c0844/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:56 crc kubenswrapper[4969]: I1004 09:55:56.454685 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b7574220-81ab-4bde-b17b-c15d3339bfd6/nova-metadata-log/0.log" Oct 04 09:55:56 crc kubenswrapper[4969]: I1004 09:55:56.834252 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_82fe815c-3f06-499f-9e3c-5a3b2f00d932/mysql-bootstrap/0.log" Oct 04 09:55:56 crc kubenswrapper[4969]: I1004 09:55:56.945837 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_9ef1ee4f-398a-49d8-a569-2d5bcc4cf073/nova-scheduler-scheduler/0.log" Oct 04 09:55:57 crc kubenswrapper[4969]: I1004 09:55:57.044197 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_82fe815c-3f06-499f-9e3c-5a3b2f00d932/galera/0.log" Oct 04 09:55:57 crc kubenswrapper[4969]: I1004 09:55:57.055075 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:55:57 crc kubenswrapper[4969]: E1004 09:55:57.055408 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:55:57 crc kubenswrapper[4969]: I1004 09:55:57.062541 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_82fe815c-3f06-499f-9e3c-5a3b2f00d932/mysql-bootstrap/0.log" Oct 04 09:55:57 crc kubenswrapper[4969]: I1004 09:55:57.282036 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8a3d9463-8792-487d-b4a5-1b61ab3310f6/mysql-bootstrap/0.log" Oct 04 09:55:57 crc kubenswrapper[4969]: I1004 09:55:57.457710 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8a3d9463-8792-487d-b4a5-1b61ab3310f6/mysql-bootstrap/0.log" Oct 04 09:55:57 crc kubenswrapper[4969]: I1004 09:55:57.473260 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8a3d9463-8792-487d-b4a5-1b61ab3310f6/galera/0.log" Oct 04 09:55:57 crc kubenswrapper[4969]: I1004 09:55:57.640733 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f5532f51-e305-4741-8661-25029faf98f2/openstackclient/0.log" Oct 04 09:55:57 crc kubenswrapper[4969]: I1004 09:55:57.715402 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-fms7j_452b4c0a-fcb1-48ed-8c3a-87c34904a0a5/ovn-controller/0.log" Oct 04 09:55:57 crc kubenswrapper[4969]: I1004 09:55:57.851552 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-zssjt_0b09ae48-946d-458d-a119-194f28bd1081/openstack-network-exporter/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.055268 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p9gfr_3f6d0bb5-43f9-4c31-838b-f71927d68b08/ovsdb-server-init/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.092015 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b7574220-81ab-4bde-b17b-c15d3339bfd6/nova-metadata-metadata/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.220899 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p9gfr_3f6d0bb5-43f9-4c31-838b-f71927d68b08/ovsdb-server-init/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.241852 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p9gfr_3f6d0bb5-43f9-4c31-838b-f71927d68b08/ovsdb-server/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.434997 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p9gfr_3f6d0bb5-43f9-4c31-838b-f71927d68b08/ovs-vswitchd/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.477216 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4hrwl_ceaa1102-1d9c-40ec-a6a7-b31ca499c55e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.555485 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_138cab2b-aabb-4736-a566-d9aee97753c0/openstack-network-exporter/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.633017 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_138cab2b-aabb-4736-a566-d9aee97753c0/ovn-northd/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.696572 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8d2f95cc-1c69-4731-90a9-a57b693b87ac/openstack-network-exporter/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.747015 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8d2f95cc-1c69-4731-90a9-a57b693b87ac/ovsdbserver-nb/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.879884 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d179b5df-13f6-4e3f-bbd5-d210ba6b2f88/openstack-network-exporter/0.log" Oct 04 09:55:58 crc kubenswrapper[4969]: I1004 09:55:58.907784 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d179b5df-13f6-4e3f-bbd5-d210ba6b2f88/ovsdbserver-sb/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.179731 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-569b746d9d-r7c45_d6425541-fb00-4545-a4df-64aba5d7e5ab/placement-api/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.196787 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/init-config-reloader/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.293956 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-569b746d9d-r7c45_d6425541-fb00-4545-a4df-64aba5d7e5ab/placement-log/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.350603 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/init-config-reloader/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.387617 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/config-reloader/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.411655 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/prometheus/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.478470 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/thanos-sidecar/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.539783 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1fdac808-4983-49ff-9975-ebc0ad82bb27/setup-container/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.732467 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1fdac808-4983-49ff-9975-ebc0ad82bb27/rabbitmq/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.760108 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1fdac808-4983-49ff-9975-ebc0ad82bb27/setup-container/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.800804 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_28264a08-8168-419d-a20d-8f94a5c4ed77/setup-container/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.939415 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_28264a08-8168-419d-a20d-8f94a5c4ed77/setup-container/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.941826 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_28264a08-8168-419d-a20d-8f94a5c4ed77/rabbitmq/0.log" Oct 04 09:55:59 crc kubenswrapper[4969]: I1004 09:55:59.984854 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7379ac07-e6f6-4218-85c2-76883f50cf0e/setup-container/0.log" Oct 04 09:56:00 crc kubenswrapper[4969]: I1004 09:56:00.149303 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7379ac07-e6f6-4218-85c2-76883f50cf0e/setup-container/0.log" Oct 04 09:56:00 crc kubenswrapper[4969]: I1004 09:56:00.168914 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7379ac07-e6f6-4218-85c2-76883f50cf0e/rabbitmq/0.log" Oct 04 09:56:00 crc kubenswrapper[4969]: I1004 09:56:00.248076 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm_942f1580-e8dc-4f73-aa37-e5a2cde068c6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:56:00 crc kubenswrapper[4969]: I1004 09:56:00.343772 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-j9fsg_562cc378-72b8-42ea-9822-4a5aa6cebbb6/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:56:00 crc kubenswrapper[4969]: I1004 09:56:00.564743 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z_a4b37224-1961-43f0-bbc9-efbc635bc063/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:56:00 crc kubenswrapper[4969]: I1004 09:56:00.699527 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kclpg_57aac488-900c-47d7-abc5-0349f1004655/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:56:00 crc kubenswrapper[4969]: I1004 09:56:00.790831 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dhdzp_b0fface3-5e9a-4a86-9863-8c7ad3c21151/ssh-known-hosts-edpm-deployment/0.log" Oct 04 09:56:00 crc kubenswrapper[4969]: I1004 09:56:00.998486 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-75f87ddd45-2md8q_f2887852-0cd7-476e-8c4e-1ed98c66ede4/proxy-server/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.028452 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-75f87ddd45-2md8q_f2887852-0cd7-476e-8c4e-1ed98c66ede4/proxy-httpd/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.147369 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-b822c_67329c10-3dda-4a99-a964-cc8039752729/swift-ring-rebalance/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.243136 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/account-auditor/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.246874 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/account-reaper/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.346447 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/account-replicator/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.395533 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/account-server/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.439972 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/container-auditor/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.459589 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/container-replicator/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.506632 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/container-server/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.619351 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/container-updater/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.627135 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-expirer/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.645296 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-auditor/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.720627 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-replicator/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.768501 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-server/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.810495 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-updater/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.814993 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/rsync/0.log" Oct 04 09:56:01 crc kubenswrapper[4969]: I1004 09:56:01.893619 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/swift-recon-cron/0.log" Oct 04 09:56:02 crc kubenswrapper[4969]: I1004 09:56:02.002032 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-k45fp_eefed526-feaa-4158-94be-19220b2801e2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:56:02 crc kubenswrapper[4969]: I1004 09:56:02.091687 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_6348f782-6e3b-4c67-b252-2276628c14d6/tempest-tests-tempest-tests-runner/0.log" Oct 04 09:56:02 crc kubenswrapper[4969]: I1004 09:56:02.168541 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_9e320e23-ff99-4d9f-b2aa-25c7d2420b88/test-operator-logs-container/0.log" Oct 04 09:56:02 crc kubenswrapper[4969]: I1004 09:56:02.259070 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7_8c7813c0-2dcd-4d30-9f8c-fac53077417f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 09:56:03 crc kubenswrapper[4969]: I1004 09:56:03.100711 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac/watcher-applier/0.log" Oct 04 09:56:03 crc kubenswrapper[4969]: I1004 09:56:03.490146 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_b36dc326-a420-42e1-a274-9e82641ee148/watcher-api-log/0.log" Oct 04 09:56:05 crc kubenswrapper[4969]: I1004 09:56:05.823457 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_6f21dc45-2a75-427d-a2c1-632c0fa0428c/watcher-decision-engine/0.log" Oct 04 09:56:06 crc kubenswrapper[4969]: I1004 09:56:06.783126 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_b36dc326-a420-42e1-a274-9e82641ee148/watcher-api/0.log" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.127363 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vljxz"] Oct 04 09:56:10 crc kubenswrapper[4969]: E1004 09:56:10.128010 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerName="extract-content" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.128023 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerName="extract-content" Oct 04 09:56:10 crc kubenswrapper[4969]: E1004 09:56:10.128061 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerName="extract-utilities" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.128067 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerName="extract-utilities" Oct 04 09:56:10 crc kubenswrapper[4969]: E1004 09:56:10.128077 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerName="registry-server" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.128083 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerName="registry-server" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.128286 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="6396e6f8-d2a1-45bd-aae3-cef656c0ef0d" containerName="registry-server" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.129809 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.136759 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vljxz"] Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.252766 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-catalog-content\") pod \"redhat-operators-vljxz\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.252823 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-utilities\") pod \"redhat-operators-vljxz\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.253090 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzfmv\" (UniqueName: \"kubernetes.io/projected/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-kube-api-access-wzfmv\") pod \"redhat-operators-vljxz\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.354764 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-catalog-content\") pod \"redhat-operators-vljxz\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.354837 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-utilities\") pod \"redhat-operators-vljxz\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.354899 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzfmv\" (UniqueName: \"kubernetes.io/projected/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-kube-api-access-wzfmv\") pod \"redhat-operators-vljxz\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.355289 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-catalog-content\") pod \"redhat-operators-vljxz\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.355652 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-utilities\") pod \"redhat-operators-vljxz\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.373844 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzfmv\" (UniqueName: \"kubernetes.io/projected/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-kube-api-access-wzfmv\") pod \"redhat-operators-vljxz\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.479311 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:10 crc kubenswrapper[4969]: I1004 09:56:10.965992 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vljxz"] Oct 04 09:56:11 crc kubenswrapper[4969]: I1004 09:56:11.055892 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:56:11 crc kubenswrapper[4969]: E1004 09:56:11.056122 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:56:11 crc kubenswrapper[4969]: I1004 09:56:11.457978 4969 generic.go:334] "Generic (PLEG): container finished" podID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerID="1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a" exitCode=0 Oct 04 09:56:11 crc kubenswrapper[4969]: I1004 09:56:11.458084 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljxz" event={"ID":"020c0fea-c12a-4cf4-b1ea-eaa6511832e5","Type":"ContainerDied","Data":"1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a"} Oct 04 09:56:11 crc kubenswrapper[4969]: I1004 09:56:11.458274 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljxz" event={"ID":"020c0fea-c12a-4cf4-b1ea-eaa6511832e5","Type":"ContainerStarted","Data":"6fa624a996a580373fa21cda1802ae83a1229c68f0bb3ed49fbbeea4edeee468"} Oct 04 09:56:13 crc kubenswrapper[4969]: I1004 09:56:13.482082 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljxz" event={"ID":"020c0fea-c12a-4cf4-b1ea-eaa6511832e5","Type":"ContainerStarted","Data":"ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9"} Oct 04 09:56:17 crc kubenswrapper[4969]: I1004 09:56:17.541457 4969 generic.go:334] "Generic (PLEG): container finished" podID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerID="ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9" exitCode=0 Oct 04 09:56:17 crc kubenswrapper[4969]: I1004 09:56:17.541881 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljxz" event={"ID":"020c0fea-c12a-4cf4-b1ea-eaa6511832e5","Type":"ContainerDied","Data":"ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9"} Oct 04 09:56:18 crc kubenswrapper[4969]: I1004 09:56:18.567855 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljxz" event={"ID":"020c0fea-c12a-4cf4-b1ea-eaa6511832e5","Type":"ContainerStarted","Data":"1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c"} Oct 04 09:56:18 crc kubenswrapper[4969]: I1004 09:56:18.585941 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vljxz" podStartSLOduration=2.015675461 podStartE2EDuration="8.585924446s" podCreationTimestamp="2025-10-04 09:56:10 +0000 UTC" firstStartedPulling="2025-10-04 09:56:11.459872889 +0000 UTC m=+5999.214141703" lastFinishedPulling="2025-10-04 09:56:18.030121874 +0000 UTC m=+6005.784390688" observedRunningTime="2025-10-04 09:56:18.582868131 +0000 UTC m=+6006.337136945" watchObservedRunningTime="2025-10-04 09:56:18.585924446 +0000 UTC m=+6006.340193260" Oct 04 09:56:20 crc kubenswrapper[4969]: I1004 09:56:20.480088 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:20 crc kubenswrapper[4969]: I1004 09:56:20.480530 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:21 crc kubenswrapper[4969]: I1004 09:56:21.537222 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vljxz" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerName="registry-server" probeResult="failure" output=< Oct 04 09:56:21 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 09:56:21 crc kubenswrapper[4969]: > Oct 04 09:56:25 crc kubenswrapper[4969]: I1004 09:56:25.059210 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:56:25 crc kubenswrapper[4969]: E1004 09:56:25.060287 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:56:30 crc kubenswrapper[4969]: I1004 09:56:30.546599 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:30 crc kubenswrapper[4969]: I1004 09:56:30.624343 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:30 crc kubenswrapper[4969]: I1004 09:56:30.790853 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vljxz"] Oct 04 09:56:31 crc kubenswrapper[4969]: I1004 09:56:31.703646 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vljxz" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerName="registry-server" containerID="cri-o://1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c" gracePeriod=2 Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.213983 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.413645 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzfmv\" (UniqueName: \"kubernetes.io/projected/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-kube-api-access-wzfmv\") pod \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.413737 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-catalog-content\") pod \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.413986 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-utilities\") pod \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\" (UID: \"020c0fea-c12a-4cf4-b1ea-eaa6511832e5\") " Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.414852 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-utilities" (OuterVolumeSpecName: "utilities") pod "020c0fea-c12a-4cf4-b1ea-eaa6511832e5" (UID: "020c0fea-c12a-4cf4-b1ea-eaa6511832e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.426862 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-kube-api-access-wzfmv" (OuterVolumeSpecName: "kube-api-access-wzfmv") pod "020c0fea-c12a-4cf4-b1ea-eaa6511832e5" (UID: "020c0fea-c12a-4cf4-b1ea-eaa6511832e5"). InnerVolumeSpecName "kube-api-access-wzfmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.489314 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "020c0fea-c12a-4cf4-b1ea-eaa6511832e5" (UID: "020c0fea-c12a-4cf4-b1ea-eaa6511832e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.515779 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.516680 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzfmv\" (UniqueName: \"kubernetes.io/projected/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-kube-api-access-wzfmv\") on node \"crc\" DevicePath \"\"" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.516706 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/020c0fea-c12a-4cf4-b1ea-eaa6511832e5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.723219 4969 generic.go:334] "Generic (PLEG): container finished" podID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerID="1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c" exitCode=0 Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.723283 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljxz" event={"ID":"020c0fea-c12a-4cf4-b1ea-eaa6511832e5","Type":"ContainerDied","Data":"1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c"} Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.723320 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljxz" event={"ID":"020c0fea-c12a-4cf4-b1ea-eaa6511832e5","Type":"ContainerDied","Data":"6fa624a996a580373fa21cda1802ae83a1229c68f0bb3ed49fbbeea4edeee468"} Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.723345 4969 scope.go:117] "RemoveContainer" containerID="1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.723283 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vljxz" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.778756 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vljxz"] Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.781325 4969 scope.go:117] "RemoveContainer" containerID="ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.791729 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vljxz"] Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.816851 4969 scope.go:117] "RemoveContainer" containerID="1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.872568 4969 scope.go:117] "RemoveContainer" containerID="1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c" Oct 04 09:56:32 crc kubenswrapper[4969]: E1004 09:56:32.873276 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c\": container with ID starting with 1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c not found: ID does not exist" containerID="1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.873316 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c"} err="failed to get container status \"1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c\": rpc error: code = NotFound desc = could not find container \"1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c\": container with ID starting with 1726c6a4b599745a2111a913178aae42a399d52d2907711b1f3d16445d04534c not found: ID does not exist" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.873351 4969 scope.go:117] "RemoveContainer" containerID="ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9" Oct 04 09:56:32 crc kubenswrapper[4969]: E1004 09:56:32.873915 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9\": container with ID starting with ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9 not found: ID does not exist" containerID="ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.873958 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9"} err="failed to get container status \"ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9\": rpc error: code = NotFound desc = could not find container \"ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9\": container with ID starting with ca625cb1afa64edda063be34d29494c515926d8e447779bd476b8771ad298dc9 not found: ID does not exist" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.873988 4969 scope.go:117] "RemoveContainer" containerID="1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a" Oct 04 09:56:32 crc kubenswrapper[4969]: E1004 09:56:32.874444 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a\": container with ID starting with 1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a not found: ID does not exist" containerID="1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a" Oct 04 09:56:32 crc kubenswrapper[4969]: I1004 09:56:32.874483 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a"} err="failed to get container status \"1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a\": rpc error: code = NotFound desc = could not find container \"1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a\": container with ID starting with 1f716791d601f8e44a6de67e9aa0bf9e9d9ee019b0da3d2ca983a0e8c17e3c3a not found: ID does not exist" Oct 04 09:56:33 crc kubenswrapper[4969]: I1004 09:56:33.093348 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" path="/var/lib/kubelet/pods/020c0fea-c12a-4cf4-b1ea-eaa6511832e5/volumes" Oct 04 09:56:40 crc kubenswrapper[4969]: I1004 09:56:40.060229 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:56:40 crc kubenswrapper[4969]: E1004 09:56:40.061377 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:56:53 crc kubenswrapper[4969]: I1004 09:56:53.069483 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:56:53 crc kubenswrapper[4969]: E1004 09:56:53.070578 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:56:56 crc kubenswrapper[4969]: I1004 09:56:56.072883 4969 generic.go:334] "Generic (PLEG): container finished" podID="3588eff0-a2f0-4233-9f16-a537bef351c3" containerID="f1110b04e100376a6109608c535768757d1c9e86735425c3f6e92dd519265321" exitCode=0 Oct 04 09:56:56 crc kubenswrapper[4969]: I1004 09:56:56.072959 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/crc-debug-xprp7" event={"ID":"3588eff0-a2f0-4233-9f16-a537bef351c3","Type":"ContainerDied","Data":"f1110b04e100376a6109608c535768757d1c9e86735425c3f6e92dd519265321"} Oct 04 09:56:57 crc kubenswrapper[4969]: I1004 09:56:57.219393 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:56:57 crc kubenswrapper[4969]: I1004 09:56:57.268417 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mtfk4/crc-debug-xprp7"] Oct 04 09:56:57 crc kubenswrapper[4969]: I1004 09:56:57.283234 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mtfk4/crc-debug-xprp7"] Oct 04 09:56:57 crc kubenswrapper[4969]: I1004 09:56:57.371674 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbqrw\" (UniqueName: \"kubernetes.io/projected/3588eff0-a2f0-4233-9f16-a537bef351c3-kube-api-access-pbqrw\") pod \"3588eff0-a2f0-4233-9f16-a537bef351c3\" (UID: \"3588eff0-a2f0-4233-9f16-a537bef351c3\") " Oct 04 09:56:57 crc kubenswrapper[4969]: I1004 09:56:57.371763 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3588eff0-a2f0-4233-9f16-a537bef351c3-host\") pod \"3588eff0-a2f0-4233-9f16-a537bef351c3\" (UID: \"3588eff0-a2f0-4233-9f16-a537bef351c3\") " Oct 04 09:56:57 crc kubenswrapper[4969]: I1004 09:56:57.372065 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3588eff0-a2f0-4233-9f16-a537bef351c3-host" (OuterVolumeSpecName: "host") pod "3588eff0-a2f0-4233-9f16-a537bef351c3" (UID: "3588eff0-a2f0-4233-9f16-a537bef351c3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 09:56:57 crc kubenswrapper[4969]: I1004 09:56:57.373054 4969 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3588eff0-a2f0-4233-9f16-a537bef351c3-host\") on node \"crc\" DevicePath \"\"" Oct 04 09:56:57 crc kubenswrapper[4969]: I1004 09:56:57.379934 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3588eff0-a2f0-4233-9f16-a537bef351c3-kube-api-access-pbqrw" (OuterVolumeSpecName: "kube-api-access-pbqrw") pod "3588eff0-a2f0-4233-9f16-a537bef351c3" (UID: "3588eff0-a2f0-4233-9f16-a537bef351c3"). InnerVolumeSpecName "kube-api-access-pbqrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:56:57 crc kubenswrapper[4969]: I1004 09:56:57.474751 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbqrw\" (UniqueName: \"kubernetes.io/projected/3588eff0-a2f0-4233-9f16-a537bef351c3-kube-api-access-pbqrw\") on node \"crc\" DevicePath \"\"" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.105447 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6882ef7a47ed2d21d144ca1ca8567c08cb1bd0c794b5d10b916ba24eff20927" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.105534 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-xprp7" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.448137 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mtfk4/crc-debug-tmfxj"] Oct 04 09:56:58 crc kubenswrapper[4969]: E1004 09:56:58.448597 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerName="registry-server" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.448611 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerName="registry-server" Oct 04 09:56:58 crc kubenswrapper[4969]: E1004 09:56:58.448632 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3588eff0-a2f0-4233-9f16-a537bef351c3" containerName="container-00" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.448638 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="3588eff0-a2f0-4233-9f16-a537bef351c3" containerName="container-00" Oct 04 09:56:58 crc kubenswrapper[4969]: E1004 09:56:58.448656 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerName="extract-utilities" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.448663 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerName="extract-utilities" Oct 04 09:56:58 crc kubenswrapper[4969]: E1004 09:56:58.448673 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerName="extract-content" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.448679 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerName="extract-content" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.448902 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="020c0fea-c12a-4cf4-b1ea-eaa6511832e5" containerName="registry-server" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.448929 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="3588eff0-a2f0-4233-9f16-a537bef351c3" containerName="container-00" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.449563 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.598174 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-host\") pod \"crc-debug-tmfxj\" (UID: \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\") " pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.598240 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv2ss\" (UniqueName: \"kubernetes.io/projected/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-kube-api-access-bv2ss\") pod \"crc-debug-tmfxj\" (UID: \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\") " pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.699731 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-host\") pod \"crc-debug-tmfxj\" (UID: \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\") " pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.699832 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv2ss\" (UniqueName: \"kubernetes.io/projected/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-kube-api-access-bv2ss\") pod \"crc-debug-tmfxj\" (UID: \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\") " pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.699846 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-host\") pod \"crc-debug-tmfxj\" (UID: \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\") " pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.730025 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv2ss\" (UniqueName: \"kubernetes.io/projected/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-kube-api-access-bv2ss\") pod \"crc-debug-tmfxj\" (UID: \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\") " pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:56:58 crc kubenswrapper[4969]: I1004 09:56:58.771202 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:56:59 crc kubenswrapper[4969]: I1004 09:56:59.078763 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3588eff0-a2f0-4233-9f16-a537bef351c3" path="/var/lib/kubelet/pods/3588eff0-a2f0-4233-9f16-a537bef351c3/volumes" Oct 04 09:56:59 crc kubenswrapper[4969]: I1004 09:56:59.121165 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" event={"ID":"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad","Type":"ContainerStarted","Data":"863c54823502497d8c50f93de4e3fdd24366712e00070fd56ded1e109581bdee"} Oct 04 09:56:59 crc kubenswrapper[4969]: I1004 09:56:59.121204 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" event={"ID":"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad","Type":"ContainerStarted","Data":"611c8cbb1d2024afe0a0eae982be5f47e9edab25bf2cc98a0691ba7fc3e8925e"} Oct 04 09:56:59 crc kubenswrapper[4969]: I1004 09:56:59.137979 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" podStartSLOduration=1.137956391 podStartE2EDuration="1.137956391s" podCreationTimestamp="2025-10-04 09:56:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 09:56:59.132333733 +0000 UTC m=+6046.886602557" watchObservedRunningTime="2025-10-04 09:56:59.137956391 +0000 UTC m=+6046.892225215" Oct 04 09:57:00 crc kubenswrapper[4969]: I1004 09:57:00.133162 4969 generic.go:334] "Generic (PLEG): container finished" podID="5fa69a9d-d435-4cb7-b51d-2d31d96d6cad" containerID="863c54823502497d8c50f93de4e3fdd24366712e00070fd56ded1e109581bdee" exitCode=0 Oct 04 09:57:00 crc kubenswrapper[4969]: I1004 09:57:00.133199 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" event={"ID":"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad","Type":"ContainerDied","Data":"863c54823502497d8c50f93de4e3fdd24366712e00070fd56ded1e109581bdee"} Oct 04 09:57:01 crc kubenswrapper[4969]: I1004 09:57:01.258200 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:57:01 crc kubenswrapper[4969]: I1004 09:57:01.341576 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv2ss\" (UniqueName: \"kubernetes.io/projected/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-kube-api-access-bv2ss\") pod \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\" (UID: \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\") " Oct 04 09:57:01 crc kubenswrapper[4969]: I1004 09:57:01.341884 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-host\") pod \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\" (UID: \"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad\") " Oct 04 09:57:01 crc kubenswrapper[4969]: I1004 09:57:01.342411 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-host" (OuterVolumeSpecName: "host") pod "5fa69a9d-d435-4cb7-b51d-2d31d96d6cad" (UID: "5fa69a9d-d435-4cb7-b51d-2d31d96d6cad"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 09:57:01 crc kubenswrapper[4969]: I1004 09:57:01.342721 4969 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-host\") on node \"crc\" DevicePath \"\"" Oct 04 09:57:01 crc kubenswrapper[4969]: I1004 09:57:01.347335 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-kube-api-access-bv2ss" (OuterVolumeSpecName: "kube-api-access-bv2ss") pod "5fa69a9d-d435-4cb7-b51d-2d31d96d6cad" (UID: "5fa69a9d-d435-4cb7-b51d-2d31d96d6cad"). InnerVolumeSpecName "kube-api-access-bv2ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:57:01 crc kubenswrapper[4969]: I1004 09:57:01.445300 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv2ss\" (UniqueName: \"kubernetes.io/projected/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad-kube-api-access-bv2ss\") on node \"crc\" DevicePath \"\"" Oct 04 09:57:02 crc kubenswrapper[4969]: I1004 09:57:02.153895 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" event={"ID":"5fa69a9d-d435-4cb7-b51d-2d31d96d6cad","Type":"ContainerDied","Data":"611c8cbb1d2024afe0a0eae982be5f47e9edab25bf2cc98a0691ba7fc3e8925e"} Oct 04 09:57:02 crc kubenswrapper[4969]: I1004 09:57:02.154198 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="611c8cbb1d2024afe0a0eae982be5f47e9edab25bf2cc98a0691ba7fc3e8925e" Oct 04 09:57:02 crc kubenswrapper[4969]: I1004 09:57:02.153953 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-tmfxj" Oct 04 09:57:06 crc kubenswrapper[4969]: I1004 09:57:06.054816 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:57:06 crc kubenswrapper[4969]: E1004 09:57:06.056478 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:57:08 crc kubenswrapper[4969]: I1004 09:57:08.713605 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mtfk4/crc-debug-tmfxj"] Oct 04 09:57:08 crc kubenswrapper[4969]: I1004 09:57:08.722482 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mtfk4/crc-debug-tmfxj"] Oct 04 09:57:09 crc kubenswrapper[4969]: I1004 09:57:09.079343 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fa69a9d-d435-4cb7-b51d-2d31d96d6cad" path="/var/lib/kubelet/pods/5fa69a9d-d435-4cb7-b51d-2d31d96d6cad/volumes" Oct 04 09:57:09 crc kubenswrapper[4969]: I1004 09:57:09.898934 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mtfk4/crc-debug-pbmh5"] Oct 04 09:57:09 crc kubenswrapper[4969]: E1004 09:57:09.900133 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa69a9d-d435-4cb7-b51d-2d31d96d6cad" containerName="container-00" Oct 04 09:57:09 crc kubenswrapper[4969]: I1004 09:57:09.900165 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa69a9d-d435-4cb7-b51d-2d31d96d6cad" containerName="container-00" Oct 04 09:57:09 crc kubenswrapper[4969]: I1004 09:57:09.900587 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fa69a9d-d435-4cb7-b51d-2d31d96d6cad" containerName="container-00" Oct 04 09:57:09 crc kubenswrapper[4969]: I1004 09:57:09.901812 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:09 crc kubenswrapper[4969]: I1004 09:57:09.983745 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5kv4\" (UniqueName: \"kubernetes.io/projected/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-kube-api-access-x5kv4\") pod \"crc-debug-pbmh5\" (UID: \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\") " pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:09 crc kubenswrapper[4969]: I1004 09:57:09.984152 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-host\") pod \"crc-debug-pbmh5\" (UID: \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\") " pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:10 crc kubenswrapper[4969]: I1004 09:57:10.086805 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5kv4\" (UniqueName: \"kubernetes.io/projected/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-kube-api-access-x5kv4\") pod \"crc-debug-pbmh5\" (UID: \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\") " pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:10 crc kubenswrapper[4969]: I1004 09:57:10.086981 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-host\") pod \"crc-debug-pbmh5\" (UID: \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\") " pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:10 crc kubenswrapper[4969]: I1004 09:57:10.088257 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-host\") pod \"crc-debug-pbmh5\" (UID: \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\") " pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:10 crc kubenswrapper[4969]: I1004 09:57:10.113297 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5kv4\" (UniqueName: \"kubernetes.io/projected/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-kube-api-access-x5kv4\") pod \"crc-debug-pbmh5\" (UID: \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\") " pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:10 crc kubenswrapper[4969]: I1004 09:57:10.236053 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:11 crc kubenswrapper[4969]: I1004 09:57:11.243756 4969 generic.go:334] "Generic (PLEG): container finished" podID="125f9585-9b77-4f72-bdf8-4dfd65ff09b1" containerID="9b78ae53f3a7aa031b55bfa51d639841a6e5567f43b7e1978b6f72e23bbaf2dc" exitCode=0 Oct 04 09:57:11 crc kubenswrapper[4969]: I1004 09:57:11.243893 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" event={"ID":"125f9585-9b77-4f72-bdf8-4dfd65ff09b1","Type":"ContainerDied","Data":"9b78ae53f3a7aa031b55bfa51d639841a6e5567f43b7e1978b6f72e23bbaf2dc"} Oct 04 09:57:11 crc kubenswrapper[4969]: I1004 09:57:11.244299 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" event={"ID":"125f9585-9b77-4f72-bdf8-4dfd65ff09b1","Type":"ContainerStarted","Data":"14cd6342d0fae76660a232246d5489512d94dcd73cf97b01125b9a93b669c044"} Oct 04 09:57:11 crc kubenswrapper[4969]: I1004 09:57:11.310329 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mtfk4/crc-debug-pbmh5"] Oct 04 09:57:11 crc kubenswrapper[4969]: I1004 09:57:11.325035 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mtfk4/crc-debug-pbmh5"] Oct 04 09:57:12 crc kubenswrapper[4969]: I1004 09:57:12.350962 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:12 crc kubenswrapper[4969]: I1004 09:57:12.447737 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5kv4\" (UniqueName: \"kubernetes.io/projected/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-kube-api-access-x5kv4\") pod \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\" (UID: \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\") " Oct 04 09:57:12 crc kubenswrapper[4969]: I1004 09:57:12.448004 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-host\") pod \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\" (UID: \"125f9585-9b77-4f72-bdf8-4dfd65ff09b1\") " Oct 04 09:57:12 crc kubenswrapper[4969]: I1004 09:57:12.448208 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-host" (OuterVolumeSpecName: "host") pod "125f9585-9b77-4f72-bdf8-4dfd65ff09b1" (UID: "125f9585-9b77-4f72-bdf8-4dfd65ff09b1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 09:57:12 crc kubenswrapper[4969]: I1004 09:57:12.448655 4969 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-host\") on node \"crc\" DevicePath \"\"" Oct 04 09:57:12 crc kubenswrapper[4969]: I1004 09:57:12.453303 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-kube-api-access-x5kv4" (OuterVolumeSpecName: "kube-api-access-x5kv4") pod "125f9585-9b77-4f72-bdf8-4dfd65ff09b1" (UID: "125f9585-9b77-4f72-bdf8-4dfd65ff09b1"). InnerVolumeSpecName "kube-api-access-x5kv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:57:12 crc kubenswrapper[4969]: I1004 09:57:12.551222 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5kv4\" (UniqueName: \"kubernetes.io/projected/125f9585-9b77-4f72-bdf8-4dfd65ff09b1-kube-api-access-x5kv4\") on node \"crc\" DevicePath \"\"" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.079303 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="125f9585-9b77-4f72-bdf8-4dfd65ff09b1" path="/var/lib/kubelet/pods/125f9585-9b77-4f72-bdf8-4dfd65ff09b1/volumes" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.264188 4969 scope.go:117] "RemoveContainer" containerID="9b78ae53f3a7aa031b55bfa51d639841a6e5567f43b7e1978b6f72e23bbaf2dc" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.264211 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/crc-debug-pbmh5" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.328591 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-x7vqg_eb8a54aa-2ce0-400c-a06f-cc3513b1df30/kube-rbac-proxy/0.log" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.405073 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-x7vqg_eb8a54aa-2ce0-400c-a06f-cc3513b1df30/manager/0.log" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.559046 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-d7885_d7dc7acf-3666-46fd-99cd-8447344cd10c/kube-rbac-proxy/0.log" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.614291 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-d7885_d7dc7acf-3666-46fd-99cd-8447344cd10c/manager/0.log" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.750365 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/util/0.log" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.906449 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/util/0.log" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.924917 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/pull/0.log" Oct 04 09:57:13 crc kubenswrapper[4969]: I1004 09:57:13.977081 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/pull/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.077670 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/pull/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.101287 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/util/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.120284 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/extract/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.276830 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-m6llv_b7837ab8-145d-4f08-b407-5f0325119fb1/kube-rbac-proxy/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.281405 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-m6llv_b7837ab8-145d-4f08-b407-5f0325119fb1/manager/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.348509 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-lvh4x_0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f/kube-rbac-proxy/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.547777 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-htxsn_fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea/kube-rbac-proxy/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.567321 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-lvh4x_0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f/manager/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.584524 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-htxsn_fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea/manager/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.744052 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-m8tn8_18413b49-54b1-4ac9-8225-4aa748d0e4f3/kube-rbac-proxy/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.775165 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-m8tn8_18413b49-54b1-4ac9-8225-4aa748d0e4f3/manager/0.log" Oct 04 09:57:14 crc kubenswrapper[4969]: I1004 09:57:14.900632 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-zg6hn_89f37d97-0acc-4940-bd3a-733f4e6ed592/kube-rbac-proxy/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.004068 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-x8xxq_edfa72c9-187e-4d6c-8aab-66cb36073282/kube-rbac-proxy/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.095646 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-zg6hn_89f37d97-0acc-4940-bd3a-733f4e6ed592/manager/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.137090 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-x8xxq_edfa72c9-187e-4d6c-8aab-66cb36073282/manager/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.175737 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-t95db_5b816f82-7f2c-4537-aadd-53b6bded2e14/kube-rbac-proxy/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.367029 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-t95db_5b816f82-7f2c-4537-aadd-53b6bded2e14/manager/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.410249 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-lqq2s_d4376d0e-3428-49ec-993f-48b32d4c8863/manager/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.427361 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-lqq2s_d4376d0e-3428-49ec-993f-48b32d4c8863/kube-rbac-proxy/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.556044 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t_5603909a-7f2a-4c86-8446-9fae64c02482/kube-rbac-proxy/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.598361 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t_5603909a-7f2a-4c86-8446-9fae64c02482/manager/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.729717 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-54zxr_0ce59178-f382-4436-be86-5bc4c5d4a9b0/kube-rbac-proxy/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.798063 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-54zxr_0ce59178-f382-4436-be86-5bc4c5d4a9b0/manager/0.log" Oct 04 09:57:15 crc kubenswrapper[4969]: I1004 09:57:15.811926 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-bg7zq_34735d37-aafa-4641-a951-f4d4d7fdab3b/kube-rbac-proxy/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.007518 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-bg7zq_34735d37-aafa-4641-a951-f4d4d7fdab3b/manager/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.014209 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-tf2wf_b457fc86-0c8f-4c98-9cbe-deddce25e44e/kube-rbac-proxy/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.037292 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-tf2wf_b457fc86-0c8f-4c98-9cbe-deddce25e44e/manager/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.190099 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g_24b8e975-2aeb-46ff-81e7-1b006bc37688/kube-rbac-proxy/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.191658 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g_24b8e975-2aeb-46ff-81e7-1b006bc37688/manager/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.306944 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-68dfdd7f5c-nl8g9_5523db3e-3b39-4bda-8885-898df0dcd5e7/kube-rbac-proxy/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.457640 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-76fbc76964-zng89_bccc7a17-2a0f-4fbb-80da-a1775673e40f/kube-rbac-proxy/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.571979 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-76fbc76964-zng89_bccc7a17-2a0f-4fbb-80da-a1775673e40f/operator/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.700944 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-d748d_03968d39-1ce4-4b22-875f-5b6f391a21c3/registry-server/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.854875 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-rq5fl_17114ec7-df18-4456-9e51-e93d74881841/kube-rbac-proxy/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.877595 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-rq5fl_17114ec7-df18-4456-9e51-e93d74881841/manager/0.log" Oct 04 09:57:16 crc kubenswrapper[4969]: I1004 09:57:16.932002 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-qxvcq_8e918774-159b-41dd-b320-5aaded1b8f52/kube-rbac-proxy/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.092148 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-qxvcq_8e918774-159b-41dd-b320-5aaded1b8f52/manager/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.153852 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz_0b27d59e-dad0-402c-b7db-7d0dd87bd68e/operator/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.301415 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-jznzg_5c506bb2-6e70-425c-b4ed-be10eb472389/kube-rbac-proxy/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.328972 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-jznzg_5c506bb2-6e70-425c-b4ed-be10eb472389/manager/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.382762 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-kt6j5_d882bca8-0cc7-4053-8684-82707b083709/kube-rbac-proxy/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.615967 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-ktq5d_8bf2e62a-8649-4b9e-8557-84441113bbdc/kube-rbac-proxy/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.640829 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-68dfdd7f5c-nl8g9_5523db3e-3b39-4bda-8885-898df0dcd5e7/manager/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.702917 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-ktq5d_8bf2e62a-8649-4b9e-8557-84441113bbdc/manager/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.761127 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-kt6j5_d882bca8-0cc7-4053-8684-82707b083709/manager/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.808160 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-676569d79-24njs_a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9/kube-rbac-proxy/0.log" Oct 04 09:57:17 crc kubenswrapper[4969]: I1004 09:57:17.882211 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-676569d79-24njs_a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9/manager/0.log" Oct 04 09:57:18 crc kubenswrapper[4969]: I1004 09:57:18.056205 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:57:18 crc kubenswrapper[4969]: E1004 09:57:18.056482 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:57:31 crc kubenswrapper[4969]: I1004 09:57:31.055352 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:57:31 crc kubenswrapper[4969]: E1004 09:57:31.056493 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:57:33 crc kubenswrapper[4969]: I1004 09:57:33.284368 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mtn99_d24a7bdf-1b12-49f8-b351-896f6edf2654/control-plane-machine-set-operator/0.log" Oct 04 09:57:33 crc kubenswrapper[4969]: I1004 09:57:33.481723 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-swbhw_5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8/kube-rbac-proxy/0.log" Oct 04 09:57:33 crc kubenswrapper[4969]: I1004 09:57:33.524264 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-swbhw_5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8/machine-api-operator/0.log" Oct 04 09:57:45 crc kubenswrapper[4969]: I1004 09:57:45.055044 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:57:45 crc kubenswrapper[4969]: E1004 09:57:45.055827 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:57:46 crc kubenswrapper[4969]: I1004 09:57:46.925272 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-w94hv_21c43c6c-8aa0-4558-8e9e-ba34798b6c18/cert-manager-controller/0.log" Oct 04 09:57:47 crc kubenswrapper[4969]: I1004 09:57:47.082401 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-8s797_daf3ed47-3e82-4a2a-ae17-d56aa124499e/cert-manager-cainjector/0.log" Oct 04 09:57:47 crc kubenswrapper[4969]: I1004 09:57:47.126698 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-rdw9w_f37f93ad-8593-4bf2-bb4a-7d01e9ab273c/cert-manager-webhook/0.log" Oct 04 09:58:00 crc kubenswrapper[4969]: I1004 09:58:00.055052 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:58:00 crc kubenswrapper[4969]: E1004 09:58:00.055749 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:58:00 crc kubenswrapper[4969]: I1004 09:58:00.082858 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-mnl4l_7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43/nmstate-console-plugin/0.log" Oct 04 09:58:00 crc kubenswrapper[4969]: I1004 09:58:00.242115 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5mjb6_eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9/nmstate-handler/0.log" Oct 04 09:58:00 crc kubenswrapper[4969]: I1004 09:58:00.305110 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-t54hh_af2281a5-7ed5-4d7b-b1ab-f97948b7fd59/kube-rbac-proxy/0.log" Oct 04 09:58:00 crc kubenswrapper[4969]: I1004 09:58:00.345237 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-t54hh_af2281a5-7ed5-4d7b-b1ab-f97948b7fd59/nmstate-metrics/0.log" Oct 04 09:58:00 crc kubenswrapper[4969]: I1004 09:58:00.492217 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-p5kmf_005b44cf-b29b-4c9d-aabf-684a15eaba71/nmstate-operator/0.log" Oct 04 09:58:00 crc kubenswrapper[4969]: I1004 09:58:00.542754 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-rsbrd_dcb7ec91-3c05-474d-aead-c3014ca7e646/nmstate-webhook/0.log" Oct 04 09:58:14 crc kubenswrapper[4969]: I1004 09:58:14.055379 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:58:14 crc kubenswrapper[4969]: E1004 09:58:14.056309 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:58:14 crc kubenswrapper[4969]: I1004 09:58:14.509844 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cfvrx_b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b/kube-rbac-proxy/0.log" Oct 04 09:58:14 crc kubenswrapper[4969]: I1004 09:58:14.678100 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cfvrx_b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b/controller/0.log" Oct 04 09:58:14 crc kubenswrapper[4969]: I1004 09:58:14.742729 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-frr-files/0.log" Oct 04 09:58:14 crc kubenswrapper[4969]: I1004 09:58:14.911755 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-reloader/0.log" Oct 04 09:58:14 crc kubenswrapper[4969]: I1004 09:58:14.912597 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-reloader/0.log" Oct 04 09:58:14 crc kubenswrapper[4969]: I1004 09:58:14.929158 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-frr-files/0.log" Oct 04 09:58:14 crc kubenswrapper[4969]: I1004 09:58:14.935242 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-metrics/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.090351 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-frr-files/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.120842 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-metrics/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.139969 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-reloader/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.160274 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-metrics/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.298072 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-reloader/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.313165 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-metrics/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.339573 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/controller/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.356131 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-frr-files/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.483893 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/frr-metrics/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.528088 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/kube-rbac-proxy/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.581270 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/kube-rbac-proxy-frr/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.743394 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/reloader/0.log" Oct 04 09:58:15 crc kubenswrapper[4969]: I1004 09:58:15.818814 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-pz8cc_897cdf44-f275-487d-85e6-9f47825f8b87/frr-k8s-webhook-server/0.log" Oct 04 09:58:16 crc kubenswrapper[4969]: I1004 09:58:16.008824 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-74c8fbc96b-hr67n_67b0a26f-7d5e-4988-83ba-8073480c2e2c/manager/0.log" Oct 04 09:58:16 crc kubenswrapper[4969]: I1004 09:58:16.190538 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6bd7b55ffd-6gshk_d0240b5e-b8bb-4690-a95a-f291b834c0d0/webhook-server/0.log" Oct 04 09:58:16 crc kubenswrapper[4969]: I1004 09:58:16.211044 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wtswn_7f4070ca-20a1-4c43-af82-890ef08c6fc3/kube-rbac-proxy/0.log" Oct 04 09:58:16 crc kubenswrapper[4969]: I1004 09:58:16.818550 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wtswn_7f4070ca-20a1-4c43-af82-890ef08c6fc3/speaker/0.log" Oct 04 09:58:17 crc kubenswrapper[4969]: I1004 09:58:17.175169 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/frr/0.log" Oct 04 09:58:26 crc kubenswrapper[4969]: I1004 09:58:26.055527 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:58:26 crc kubenswrapper[4969]: E1004 09:58:26.056293 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.219653 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/util/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.363218 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/util/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.410010 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/pull/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.411582 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/pull/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.588104 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/util/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.588631 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/pull/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.602563 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/extract/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.753981 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/util/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.938531 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/util/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.965943 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/pull/0.log" Oct 04 09:58:30 crc kubenswrapper[4969]: I1004 09:58:30.968733 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/pull/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.080732 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/util/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.105729 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/pull/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.154630 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/extract/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.264465 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-utilities/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.450837 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-utilities/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.471317 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-content/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.489193 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-content/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.659083 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-content/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.687624 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-utilities/0.log" Oct 04 09:58:31 crc kubenswrapper[4969]: I1004 09:58:31.885306 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-utilities/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.128266 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-utilities/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.146798 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-content/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.212567 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-content/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.378759 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-content/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.397765 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/registry-server/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.412234 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-utilities/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.642329 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/util/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.836485 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/pull/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.869867 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/pull/0.log" Oct 04 09:58:32 crc kubenswrapper[4969]: I1004 09:58:32.881576 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/util/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.070301 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/util/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.097459 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/extract/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.142055 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/pull/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.311414 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/registry-server/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.314781 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5qth5_0cf38a30-f714-4a15-8be5-30118e8984c7/marketplace-operator/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.462484 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-utilities/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.629006 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-utilities/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.652277 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-content/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.673637 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-content/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.826930 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-content/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.839883 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-utilities/0.log" Oct 04 09:58:33 crc kubenswrapper[4969]: I1004 09:58:33.854539 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-utilities/0.log" Oct 04 09:58:34 crc kubenswrapper[4969]: I1004 09:58:34.037533 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/registry-server/0.log" Oct 04 09:58:34 crc kubenswrapper[4969]: I1004 09:58:34.093557 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-content/0.log" Oct 04 09:58:34 crc kubenswrapper[4969]: I1004 09:58:34.131471 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-utilities/0.log" Oct 04 09:58:34 crc kubenswrapper[4969]: I1004 09:58:34.173287 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-content/0.log" Oct 04 09:58:34 crc kubenswrapper[4969]: I1004 09:58:34.290038 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-utilities/0.log" Oct 04 09:58:34 crc kubenswrapper[4969]: I1004 09:58:34.352283 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-content/0.log" Oct 04 09:58:35 crc kubenswrapper[4969]: I1004 09:58:35.069972 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/registry-server/0.log" Oct 04 09:58:39 crc kubenswrapper[4969]: I1004 09:58:39.056078 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:58:39 crc kubenswrapper[4969]: E1004 09:58:39.057022 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:58:47 crc kubenswrapper[4969]: I1004 09:58:47.399243 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-2xmcb_91044ecd-9c81-4052-9e16-9224f04f079e/prometheus-operator/0.log" Oct 04 09:58:47 crc kubenswrapper[4969]: I1004 09:58:47.538115 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6_fca6110d-f0fb-4b9e-8c48-749c188c96a3/prometheus-operator-admission-webhook/0.log" Oct 04 09:58:47 crc kubenswrapper[4969]: I1004 09:58:47.605791 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn_c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e/prometheus-operator-admission-webhook/0.log" Oct 04 09:58:47 crc kubenswrapper[4969]: I1004 09:58:47.733474 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-vl7q6_15b3f66e-d095-42c3-94e2-b9ee1071d638/operator/0.log" Oct 04 09:58:47 crc kubenswrapper[4969]: I1004 09:58:47.810994 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-2c6cm_46090d90-0cf0-47f4-a19f-ad5aa65df521/perses-operator/0.log" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.276322 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dxh2h"] Oct 04 09:58:48 crc kubenswrapper[4969]: E1004 09:58:48.277057 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="125f9585-9b77-4f72-bdf8-4dfd65ff09b1" containerName="container-00" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.277076 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="125f9585-9b77-4f72-bdf8-4dfd65ff09b1" containerName="container-00" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.277277 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="125f9585-9b77-4f72-bdf8-4dfd65ff09b1" containerName="container-00" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.278801 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.311502 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxh2h"] Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.450047 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62jzr\" (UniqueName: \"kubernetes.io/projected/812d6d2a-dabf-4016-b119-c13bfaca99e2-kube-api-access-62jzr\") pod \"redhat-marketplace-dxh2h\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.450139 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-utilities\") pod \"redhat-marketplace-dxh2h\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.450177 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-catalog-content\") pod \"redhat-marketplace-dxh2h\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.551440 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62jzr\" (UniqueName: \"kubernetes.io/projected/812d6d2a-dabf-4016-b119-c13bfaca99e2-kube-api-access-62jzr\") pod \"redhat-marketplace-dxh2h\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.551529 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-utilities\") pod \"redhat-marketplace-dxh2h\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.551566 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-catalog-content\") pod \"redhat-marketplace-dxh2h\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.552005 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-catalog-content\") pod \"redhat-marketplace-dxh2h\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.552108 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-utilities\") pod \"redhat-marketplace-dxh2h\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.580269 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62jzr\" (UniqueName: \"kubernetes.io/projected/812d6d2a-dabf-4016-b119-c13bfaca99e2-kube-api-access-62jzr\") pod \"redhat-marketplace-dxh2h\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:48 crc kubenswrapper[4969]: I1004 09:58:48.602187 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:49 crc kubenswrapper[4969]: I1004 09:58:49.137718 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxh2h"] Oct 04 09:58:49 crc kubenswrapper[4969]: I1004 09:58:49.321562 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxh2h" event={"ID":"812d6d2a-dabf-4016-b119-c13bfaca99e2","Type":"ContainerStarted","Data":"8fdc656bd15ee8203ed0a3e5c17413ae3de55441dcc668157204c7d248f01f05"} Oct 04 09:58:50 crc kubenswrapper[4969]: I1004 09:58:50.336310 4969 generic.go:334] "Generic (PLEG): container finished" podID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerID="510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097" exitCode=0 Oct 04 09:58:50 crc kubenswrapper[4969]: I1004 09:58:50.336408 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxh2h" event={"ID":"812d6d2a-dabf-4016-b119-c13bfaca99e2","Type":"ContainerDied","Data":"510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097"} Oct 04 09:58:50 crc kubenswrapper[4969]: I1004 09:58:50.340058 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 09:58:52 crc kubenswrapper[4969]: I1004 09:58:52.363196 4969 generic.go:334] "Generic (PLEG): container finished" podID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerID="90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c" exitCode=0 Oct 04 09:58:52 crc kubenswrapper[4969]: I1004 09:58:52.363395 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxh2h" event={"ID":"812d6d2a-dabf-4016-b119-c13bfaca99e2","Type":"ContainerDied","Data":"90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c"} Oct 04 09:58:53 crc kubenswrapper[4969]: I1004 09:58:53.065468 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:58:53 crc kubenswrapper[4969]: E1004 09:58:53.065972 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:58:53 crc kubenswrapper[4969]: I1004 09:58:53.379668 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxh2h" event={"ID":"812d6d2a-dabf-4016-b119-c13bfaca99e2","Type":"ContainerStarted","Data":"c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0"} Oct 04 09:58:53 crc kubenswrapper[4969]: I1004 09:58:53.405016 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dxh2h" podStartSLOduration=2.805310691 podStartE2EDuration="5.404999315s" podCreationTimestamp="2025-10-04 09:58:48 +0000 UTC" firstStartedPulling="2025-10-04 09:58:50.339679415 +0000 UTC m=+6158.093948269" lastFinishedPulling="2025-10-04 09:58:52.939368079 +0000 UTC m=+6160.693636893" observedRunningTime="2025-10-04 09:58:53.402096343 +0000 UTC m=+6161.156365177" watchObservedRunningTime="2025-10-04 09:58:53.404999315 +0000 UTC m=+6161.159268129" Oct 04 09:58:58 crc kubenswrapper[4969]: I1004 09:58:58.603180 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:58 crc kubenswrapper[4969]: I1004 09:58:58.603436 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:58 crc kubenswrapper[4969]: I1004 09:58:58.674912 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:59 crc kubenswrapper[4969]: I1004 09:58:59.519558 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:58:59 crc kubenswrapper[4969]: I1004 09:58:59.574655 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxh2h"] Oct 04 09:59:01 crc kubenswrapper[4969]: I1004 09:59:01.452793 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dxh2h" podUID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerName="registry-server" containerID="cri-o://c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0" gracePeriod=2 Oct 04 09:59:01 crc kubenswrapper[4969]: E1004 09:59:01.539584 4969 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.195:45506->38.102.83.195:43269: write tcp 38.102.83.195:45506->38.102.83.195:43269: write: broken pipe Oct 04 09:59:01 crc kubenswrapper[4969]: I1004 09:59:01.982839 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.135931 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-catalog-content\") pod \"812d6d2a-dabf-4016-b119-c13bfaca99e2\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.136522 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62jzr\" (UniqueName: \"kubernetes.io/projected/812d6d2a-dabf-4016-b119-c13bfaca99e2-kube-api-access-62jzr\") pod \"812d6d2a-dabf-4016-b119-c13bfaca99e2\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.136599 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-utilities\") pod \"812d6d2a-dabf-4016-b119-c13bfaca99e2\" (UID: \"812d6d2a-dabf-4016-b119-c13bfaca99e2\") " Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.137599 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-utilities" (OuterVolumeSpecName: "utilities") pod "812d6d2a-dabf-4016-b119-c13bfaca99e2" (UID: "812d6d2a-dabf-4016-b119-c13bfaca99e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.147577 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "812d6d2a-dabf-4016-b119-c13bfaca99e2" (UID: "812d6d2a-dabf-4016-b119-c13bfaca99e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.158573 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/812d6d2a-dabf-4016-b119-c13bfaca99e2-kube-api-access-62jzr" (OuterVolumeSpecName: "kube-api-access-62jzr") pod "812d6d2a-dabf-4016-b119-c13bfaca99e2" (UID: "812d6d2a-dabf-4016-b119-c13bfaca99e2"). InnerVolumeSpecName "kube-api-access-62jzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.239330 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.239362 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812d6d2a-dabf-4016-b119-c13bfaca99e2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.239373 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62jzr\" (UniqueName: \"kubernetes.io/projected/812d6d2a-dabf-4016-b119-c13bfaca99e2-kube-api-access-62jzr\") on node \"crc\" DevicePath \"\"" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.462875 4969 generic.go:334] "Generic (PLEG): container finished" podID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerID="c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0" exitCode=0 Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.462916 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxh2h" event={"ID":"812d6d2a-dabf-4016-b119-c13bfaca99e2","Type":"ContainerDied","Data":"c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0"} Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.462932 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dxh2h" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.462951 4969 scope.go:117] "RemoveContainer" containerID="c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.462940 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxh2h" event={"ID":"812d6d2a-dabf-4016-b119-c13bfaca99e2","Type":"ContainerDied","Data":"8fdc656bd15ee8203ed0a3e5c17413ae3de55441dcc668157204c7d248f01f05"} Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.491364 4969 scope.go:117] "RemoveContainer" containerID="90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.516142 4969 scope.go:117] "RemoveContainer" containerID="510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.519015 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxh2h"] Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.541061 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxh2h"] Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.569245 4969 scope.go:117] "RemoveContainer" containerID="c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0" Oct 04 09:59:02 crc kubenswrapper[4969]: E1004 09:59:02.569855 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0\": container with ID starting with c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0 not found: ID does not exist" containerID="c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.569902 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0"} err="failed to get container status \"c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0\": rpc error: code = NotFound desc = could not find container \"c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0\": container with ID starting with c5a5e5b8a04d31dd5e1b76e756164e439eb796801902f18c725d9381cda213a0 not found: ID does not exist" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.569932 4969 scope.go:117] "RemoveContainer" containerID="90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c" Oct 04 09:59:02 crc kubenswrapper[4969]: E1004 09:59:02.570353 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c\": container with ID starting with 90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c not found: ID does not exist" containerID="90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.570387 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c"} err="failed to get container status \"90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c\": rpc error: code = NotFound desc = could not find container \"90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c\": container with ID starting with 90a85ed4801d6d50674e0902df22d809791399d83edf81a10e543ef1d548438c not found: ID does not exist" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.570406 4969 scope.go:117] "RemoveContainer" containerID="510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097" Oct 04 09:59:02 crc kubenswrapper[4969]: E1004 09:59:02.570891 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097\": container with ID starting with 510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097 not found: ID does not exist" containerID="510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097" Oct 04 09:59:02 crc kubenswrapper[4969]: I1004 09:59:02.570922 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097"} err="failed to get container status \"510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097\": rpc error: code = NotFound desc = could not find container \"510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097\": container with ID starting with 510ef66c48640160a234b6f1019901e44dbe7be6b8e35eaebfdd7f97de268097 not found: ID does not exist" Oct 04 09:59:03 crc kubenswrapper[4969]: I1004 09:59:03.066297 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="812d6d2a-dabf-4016-b119-c13bfaca99e2" path="/var/lib/kubelet/pods/812d6d2a-dabf-4016-b119-c13bfaca99e2/volumes" Oct 04 09:59:07 crc kubenswrapper[4969]: I1004 09:59:07.059678 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:59:07 crc kubenswrapper[4969]: E1004 09:59:07.064861 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:59:20 crc kubenswrapper[4969]: I1004 09:59:20.056554 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:59:20 crc kubenswrapper[4969]: E1004 09:59:20.057622 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:59:34 crc kubenswrapper[4969]: I1004 09:59:34.056587 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:59:34 crc kubenswrapper[4969]: E1004 09:59:34.057261 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 09:59:48 crc kubenswrapper[4969]: I1004 09:59:48.054988 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 09:59:48 crc kubenswrapper[4969]: E1004 09:59:48.055755 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.057828 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.175214 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn"] Oct 04 10:00:00 crc kubenswrapper[4969]: E1004 10:00:00.176523 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerName="extract-content" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.176545 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerName="extract-content" Oct 04 10:00:00 crc kubenswrapper[4969]: E1004 10:00:00.176573 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerName="extract-utilities" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.176582 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerName="extract-utilities" Oct 04 10:00:00 crc kubenswrapper[4969]: E1004 10:00:00.176590 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerName="registry-server" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.176597 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerName="registry-server" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.177134 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="812d6d2a-dabf-4016-b119-c13bfaca99e2" containerName="registry-server" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.181210 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.191183 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn"] Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.191566 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.191842 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.294770 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c8cc7a-bceb-4152-9647-555b0c8c9965-secret-volume\") pod \"collect-profiles-29326200-6lqmn\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.294821 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bwkw\" (UniqueName: \"kubernetes.io/projected/a3c8cc7a-bceb-4152-9647-555b0c8c9965-kube-api-access-9bwkw\") pod \"collect-profiles-29326200-6lqmn\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.294870 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c8cc7a-bceb-4152-9647-555b0c8c9965-config-volume\") pod \"collect-profiles-29326200-6lqmn\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.397102 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bwkw\" (UniqueName: \"kubernetes.io/projected/a3c8cc7a-bceb-4152-9647-555b0c8c9965-kube-api-access-9bwkw\") pod \"collect-profiles-29326200-6lqmn\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.397237 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c8cc7a-bceb-4152-9647-555b0c8c9965-config-volume\") pod \"collect-profiles-29326200-6lqmn\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.397490 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c8cc7a-bceb-4152-9647-555b0c8c9965-secret-volume\") pod \"collect-profiles-29326200-6lqmn\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.399491 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c8cc7a-bceb-4152-9647-555b0c8c9965-config-volume\") pod \"collect-profiles-29326200-6lqmn\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.421149 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c8cc7a-bceb-4152-9647-555b0c8c9965-secret-volume\") pod \"collect-profiles-29326200-6lqmn\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.426620 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bwkw\" (UniqueName: \"kubernetes.io/projected/a3c8cc7a-bceb-4152-9647-555b0c8c9965-kube-api-access-9bwkw\") pod \"collect-profiles-29326200-6lqmn\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:00 crc kubenswrapper[4969]: I1004 10:00:00.526342 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:01 crc kubenswrapper[4969]: I1004 10:00:01.008222 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn"] Oct 04 10:00:01 crc kubenswrapper[4969]: W1004 10:00:01.015565 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3c8cc7a_bceb_4152_9647_555b0c8c9965.slice/crio-1d41f08db13591346a3e45f58e76f15a8ca0b60e7a74da69f522a9a45b4e1908 WatchSource:0}: Error finding container 1d41f08db13591346a3e45f58e76f15a8ca0b60e7a74da69f522a9a45b4e1908: Status 404 returned error can't find the container with id 1d41f08db13591346a3e45f58e76f15a8ca0b60e7a74da69f522a9a45b4e1908 Oct 04 10:00:01 crc kubenswrapper[4969]: I1004 10:00:01.231875 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"8fdf0e4cb48a4d351d146bddccf73a2a4695a141ecf2ec805971c3bfed433381"} Oct 04 10:00:01 crc kubenswrapper[4969]: I1004 10:00:01.234288 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" event={"ID":"a3c8cc7a-bceb-4152-9647-555b0c8c9965","Type":"ContainerStarted","Data":"1d41f08db13591346a3e45f58e76f15a8ca0b60e7a74da69f522a9a45b4e1908"} Oct 04 10:00:02 crc kubenswrapper[4969]: I1004 10:00:02.253047 4969 generic.go:334] "Generic (PLEG): container finished" podID="a3c8cc7a-bceb-4152-9647-555b0c8c9965" containerID="6ed527097b7f1c90bdd231b9502219a5c53165464070e405702c02ef7c6f0725" exitCode=0 Oct 04 10:00:02 crc kubenswrapper[4969]: I1004 10:00:02.253344 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" event={"ID":"a3c8cc7a-bceb-4152-9647-555b0c8c9965","Type":"ContainerDied","Data":"6ed527097b7f1c90bdd231b9502219a5c53165464070e405702c02ef7c6f0725"} Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.710118 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.870452 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bwkw\" (UniqueName: \"kubernetes.io/projected/a3c8cc7a-bceb-4152-9647-555b0c8c9965-kube-api-access-9bwkw\") pod \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.870844 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c8cc7a-bceb-4152-9647-555b0c8c9965-config-volume\") pod \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.870899 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c8cc7a-bceb-4152-9647-555b0c8c9965-secret-volume\") pod \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\" (UID: \"a3c8cc7a-bceb-4152-9647-555b0c8c9965\") " Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.872254 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3c8cc7a-bceb-4152-9647-555b0c8c9965-config-volume" (OuterVolumeSpecName: "config-volume") pod "a3c8cc7a-bceb-4152-9647-555b0c8c9965" (UID: "a3c8cc7a-bceb-4152-9647-555b0c8c9965"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.876083 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3c8cc7a-bceb-4152-9647-555b0c8c9965-kube-api-access-9bwkw" (OuterVolumeSpecName: "kube-api-access-9bwkw") pod "a3c8cc7a-bceb-4152-9647-555b0c8c9965" (UID: "a3c8cc7a-bceb-4152-9647-555b0c8c9965"). InnerVolumeSpecName "kube-api-access-9bwkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.876570 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c8cc7a-bceb-4152-9647-555b0c8c9965-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a3c8cc7a-bceb-4152-9647-555b0c8c9965" (UID: "a3c8cc7a-bceb-4152-9647-555b0c8c9965"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.973688 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bwkw\" (UniqueName: \"kubernetes.io/projected/a3c8cc7a-bceb-4152-9647-555b0c8c9965-kube-api-access-9bwkw\") on node \"crc\" DevicePath \"\"" Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.973727 4969 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c8cc7a-bceb-4152-9647-555b0c8c9965-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:00:03 crc kubenswrapper[4969]: I1004 10:00:03.973739 4969 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c8cc7a-bceb-4152-9647-555b0c8c9965-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:00:04 crc kubenswrapper[4969]: I1004 10:00:04.283962 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" event={"ID":"a3c8cc7a-bceb-4152-9647-555b0c8c9965","Type":"ContainerDied","Data":"1d41f08db13591346a3e45f58e76f15a8ca0b60e7a74da69f522a9a45b4e1908"} Oct 04 10:00:04 crc kubenswrapper[4969]: I1004 10:00:04.284062 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d41f08db13591346a3e45f58e76f15a8ca0b60e7a74da69f522a9a45b4e1908" Oct 04 10:00:04 crc kubenswrapper[4969]: I1004 10:00:04.284187 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326200-6lqmn" Oct 04 10:00:04 crc kubenswrapper[4969]: I1004 10:00:04.802903 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln"] Oct 04 10:00:04 crc kubenswrapper[4969]: I1004 10:00:04.805978 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326155-hq9ln"] Oct 04 10:00:05 crc kubenswrapper[4969]: I1004 10:00:05.073716 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6243bf2a-5d33-41f1-a09f-bd0e441fbe4e" path="/var/lib/kubelet/pods/6243bf2a-5d33-41f1-a09f-bd0e441fbe4e/volumes" Oct 04 10:00:35 crc kubenswrapper[4969]: I1004 10:00:35.130932 4969 scope.go:117] "RemoveContainer" containerID="8e364dbc4824ba12e15260a6705964c53cb2061225c512fd1833f6d786805f10" Oct 04 10:00:56 crc kubenswrapper[4969]: I1004 10:00:56.942778 4969 generic.go:334] "Generic (PLEG): container finished" podID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" containerID="0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905" exitCode=0 Oct 04 10:00:56 crc kubenswrapper[4969]: I1004 10:00:56.942966 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" event={"ID":"042bd31c-d47c-488f-b1e6-4765f3b3cbc8","Type":"ContainerDied","Data":"0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905"} Oct 04 10:00:56 crc kubenswrapper[4969]: I1004 10:00:56.944206 4969 scope.go:117] "RemoveContainer" containerID="0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905" Oct 04 10:00:57 crc kubenswrapper[4969]: I1004 10:00:57.350174 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mtfk4_must-gather-xz6qb_042bd31c-d47c-488f-b1e6-4765f3b3cbc8/gather/0.log" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.138650 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29326201-l86vc"] Oct 04 10:01:00 crc kubenswrapper[4969]: E1004 10:01:00.139533 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3c8cc7a-bceb-4152-9647-555b0c8c9965" containerName="collect-profiles" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.139548 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3c8cc7a-bceb-4152-9647-555b0c8c9965" containerName="collect-profiles" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.139762 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3c8cc7a-bceb-4152-9647-555b0c8c9965" containerName="collect-profiles" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.140388 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.161721 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29326201-l86vc"] Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.233739 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dndwt\" (UniqueName: \"kubernetes.io/projected/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-kube-api-access-dndwt\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.233972 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-combined-ca-bundle\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.233998 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-config-data\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.234024 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-fernet-keys\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.335560 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dndwt\" (UniqueName: \"kubernetes.io/projected/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-kube-api-access-dndwt\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.335759 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-combined-ca-bundle\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.337328 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-config-data\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.337380 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-fernet-keys\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.344621 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-combined-ca-bundle\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.344876 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-config-data\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.345746 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-fernet-keys\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.350141 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dndwt\" (UniqueName: \"kubernetes.io/projected/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-kube-api-access-dndwt\") pod \"keystone-cron-29326201-l86vc\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.470323 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.959021 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29326201-l86vc"] Oct 04 10:01:00 crc kubenswrapper[4969]: I1004 10:01:00.994167 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326201-l86vc" event={"ID":"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad","Type":"ContainerStarted","Data":"f9aca4c35fc0b969175bd61e9480013c33b6883adbcfe9cec467cd0eafaf098e"} Oct 04 10:01:02 crc kubenswrapper[4969]: I1004 10:01:02.006531 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326201-l86vc" event={"ID":"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad","Type":"ContainerStarted","Data":"e115f5e111d81b7c24cb21882bf6e2d133771ca7a2ca38acbf339483ee5c5325"} Oct 04 10:01:02 crc kubenswrapper[4969]: I1004 10:01:02.032185 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29326201-l86vc" podStartSLOduration=2.03212033 podStartE2EDuration="2.03212033s" podCreationTimestamp="2025-10-04 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:01:02.029980098 +0000 UTC m=+6289.784248952" watchObservedRunningTime="2025-10-04 10:01:02.03212033 +0000 UTC m=+6289.786389164" Oct 04 10:01:05 crc kubenswrapper[4969]: I1004 10:01:05.045897 4969 generic.go:334] "Generic (PLEG): container finished" podID="b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad" containerID="e115f5e111d81b7c24cb21882bf6e2d133771ca7a2ca38acbf339483ee5c5325" exitCode=0 Oct 04 10:01:05 crc kubenswrapper[4969]: I1004 10:01:05.046369 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326201-l86vc" event={"ID":"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad","Type":"ContainerDied","Data":"e115f5e111d81b7c24cb21882bf6e2d133771ca7a2ca38acbf339483ee5c5325"} Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.018014 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mtfk4/must-gather-xz6qb"] Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.018342 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" podUID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" containerName="copy" containerID="cri-o://13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b" gracePeriod=2 Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.028302 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mtfk4/must-gather-xz6qb"] Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.554358 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.560187 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mtfk4_must-gather-xz6qb_042bd31c-d47c-488f-b1e6-4765f3b3cbc8/copy/0.log" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.560665 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.689565 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-combined-ca-bundle\") pod \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.689647 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj9k5\" (UniqueName: \"kubernetes.io/projected/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-kube-api-access-cj9k5\") pod \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\" (UID: \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\") " Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.689706 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-must-gather-output\") pod \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\" (UID: \"042bd31c-d47c-488f-b1e6-4765f3b3cbc8\") " Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.689791 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-config-data\") pod \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.692878 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-fernet-keys\") pod \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.693010 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dndwt\" (UniqueName: \"kubernetes.io/projected/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-kube-api-access-dndwt\") pod \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\" (UID: \"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad\") " Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.699790 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-kube-api-access-cj9k5" (OuterVolumeSpecName: "kube-api-access-cj9k5") pod "042bd31c-d47c-488f-b1e6-4765f3b3cbc8" (UID: "042bd31c-d47c-488f-b1e6-4765f3b3cbc8"). InnerVolumeSpecName "kube-api-access-cj9k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.700025 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-kube-api-access-dndwt" (OuterVolumeSpecName: "kube-api-access-dndwt") pod "b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad" (UID: "b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad"). InnerVolumeSpecName "kube-api-access-dndwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.701565 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad" (UID: "b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.725828 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad" (UID: "b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.754219 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-config-data" (OuterVolumeSpecName: "config-data") pod "b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad" (UID: "b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.796837 4969 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.797095 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj9k5\" (UniqueName: \"kubernetes.io/projected/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-kube-api-access-cj9k5\") on node \"crc\" DevicePath \"\"" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.797105 4969 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.797113 4969 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.797122 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dndwt\" (UniqueName: \"kubernetes.io/projected/b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad-kube-api-access-dndwt\") on node \"crc\" DevicePath \"\"" Oct 04 10:01:06 crc kubenswrapper[4969]: I1004 10:01:06.902023 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "042bd31c-d47c-488f-b1e6-4765f3b3cbc8" (UID: "042bd31c-d47c-488f-b1e6-4765f3b3cbc8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.001776 4969 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/042bd31c-d47c-488f-b1e6-4765f3b3cbc8-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.068601 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" path="/var/lib/kubelet/pods/042bd31c-d47c-488f-b1e6-4765f3b3cbc8/volumes" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.081439 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mtfk4_must-gather-xz6qb_042bd31c-d47c-488f-b1e6-4765f3b3cbc8/copy/0.log" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.081852 4969 generic.go:334] "Generic (PLEG): container finished" podID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" containerID="13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b" exitCode=143 Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.081912 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtfk4/must-gather-xz6qb" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.081948 4969 scope.go:117] "RemoveContainer" containerID="13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.084081 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326201-l86vc" event={"ID":"b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad","Type":"ContainerDied","Data":"f9aca4c35fc0b969175bd61e9480013c33b6883adbcfe9cec467cd0eafaf098e"} Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.084125 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9aca4c35fc0b969175bd61e9480013c33b6883adbcfe9cec467cd0eafaf098e" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.084200 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326201-l86vc" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.113976 4969 scope.go:117] "RemoveContainer" containerID="0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.197931 4969 scope.go:117] "RemoveContainer" containerID="13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b" Oct 04 10:01:07 crc kubenswrapper[4969]: E1004 10:01:07.198359 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b\": container with ID starting with 13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b not found: ID does not exist" containerID="13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.198389 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b"} err="failed to get container status \"13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b\": rpc error: code = NotFound desc = could not find container \"13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b\": container with ID starting with 13a377e19eb54305f3bfaa455c35fcfe293001ed8bf2c20ac57bad15561c526b not found: ID does not exist" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.198410 4969 scope.go:117] "RemoveContainer" containerID="0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905" Oct 04 10:01:07 crc kubenswrapper[4969]: E1004 10:01:07.198594 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905\": container with ID starting with 0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905 not found: ID does not exist" containerID="0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905" Oct 04 10:01:07 crc kubenswrapper[4969]: I1004 10:01:07.198608 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905"} err="failed to get container status \"0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905\": rpc error: code = NotFound desc = could not find container \"0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905\": container with ID starting with 0dd599b86a3bf109edc431193c968e67e5b1a8f5778101acd925de8c3de23905 not found: ID does not exist" Oct 04 10:01:35 crc kubenswrapper[4969]: I1004 10:01:35.254862 4969 scope.go:117] "RemoveContainer" containerID="f1110b04e100376a6109608c535768757d1c9e86735425c3f6e92dd519265321" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.944509 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6b4sb/must-gather-jbgx9"] Oct 04 10:01:45 crc kubenswrapper[4969]: E1004 10:01:45.945296 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" containerName="copy" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.945309 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" containerName="copy" Oct 04 10:01:45 crc kubenswrapper[4969]: E1004 10:01:45.945319 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" containerName="gather" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.945324 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" containerName="gather" Oct 04 10:01:45 crc kubenswrapper[4969]: E1004 10:01:45.945332 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad" containerName="keystone-cron" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.945338 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad" containerName="keystone-cron" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.945571 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" containerName="copy" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.945590 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad" containerName="keystone-cron" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.945601 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="042bd31c-d47c-488f-b1e6-4765f3b3cbc8" containerName="gather" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.946601 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.948242 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6b4sb"/"openshift-service-ca.crt" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.949135 4969 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-6b4sb"/"default-dockercfg-69tw6" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.951170 4969 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6b4sb"/"kube-root-ca.crt" Oct 04 10:01:45 crc kubenswrapper[4969]: I1004 10:01:45.963240 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6b4sb/must-gather-jbgx9"] Oct 04 10:01:46 crc kubenswrapper[4969]: I1004 10:01:46.015703 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cbdba948-d8b5-4828-b27d-d9a93a49717a-must-gather-output\") pod \"must-gather-jbgx9\" (UID: \"cbdba948-d8b5-4828-b27d-d9a93a49717a\") " pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:01:46 crc kubenswrapper[4969]: I1004 10:01:46.015764 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clfd8\" (UniqueName: \"kubernetes.io/projected/cbdba948-d8b5-4828-b27d-d9a93a49717a-kube-api-access-clfd8\") pod \"must-gather-jbgx9\" (UID: \"cbdba948-d8b5-4828-b27d-d9a93a49717a\") " pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:01:46 crc kubenswrapper[4969]: I1004 10:01:46.158865 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cbdba948-d8b5-4828-b27d-d9a93a49717a-must-gather-output\") pod \"must-gather-jbgx9\" (UID: \"cbdba948-d8b5-4828-b27d-d9a93a49717a\") " pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:01:46 crc kubenswrapper[4969]: I1004 10:01:46.159010 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clfd8\" (UniqueName: \"kubernetes.io/projected/cbdba948-d8b5-4828-b27d-d9a93a49717a-kube-api-access-clfd8\") pod \"must-gather-jbgx9\" (UID: \"cbdba948-d8b5-4828-b27d-d9a93a49717a\") " pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:01:46 crc kubenswrapper[4969]: I1004 10:01:46.159966 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cbdba948-d8b5-4828-b27d-d9a93a49717a-must-gather-output\") pod \"must-gather-jbgx9\" (UID: \"cbdba948-d8b5-4828-b27d-d9a93a49717a\") " pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:01:46 crc kubenswrapper[4969]: I1004 10:01:46.177898 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clfd8\" (UniqueName: \"kubernetes.io/projected/cbdba948-d8b5-4828-b27d-d9a93a49717a-kube-api-access-clfd8\") pod \"must-gather-jbgx9\" (UID: \"cbdba948-d8b5-4828-b27d-d9a93a49717a\") " pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:01:46 crc kubenswrapper[4969]: I1004 10:01:46.266646 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:01:46 crc kubenswrapper[4969]: I1004 10:01:46.754110 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6b4sb/must-gather-jbgx9"] Oct 04 10:01:47 crc kubenswrapper[4969]: I1004 10:01:47.549983 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" event={"ID":"cbdba948-d8b5-4828-b27d-d9a93a49717a","Type":"ContainerStarted","Data":"338011150e91a2246896914e7107c5f2e91f5523aa3a99007693cf04675f48d4"} Oct 04 10:01:47 crc kubenswrapper[4969]: I1004 10:01:47.550026 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" event={"ID":"cbdba948-d8b5-4828-b27d-d9a93a49717a","Type":"ContainerStarted","Data":"f0b0ffae271acfc0b781ee4ace2439b173ab1ff54b9a2f1eb663a69b67d1dd05"} Oct 04 10:01:47 crc kubenswrapper[4969]: I1004 10:01:47.550037 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" event={"ID":"cbdba948-d8b5-4828-b27d-d9a93a49717a","Type":"ContainerStarted","Data":"33e2ad8fb38464e9c86c5c20e1ebcc917b351788c4ff23441e4f1f977b36d546"} Oct 04 10:01:47 crc kubenswrapper[4969]: I1004 10:01:47.573143 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" podStartSLOduration=2.573120527 podStartE2EDuration="2.573120527s" podCreationTimestamp="2025-10-04 10:01:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:01:47.571794864 +0000 UTC m=+6335.326063678" watchObservedRunningTime="2025-10-04 10:01:47.573120527 +0000 UTC m=+6335.327389341" Oct 04 10:01:50 crc kubenswrapper[4969]: I1004 10:01:50.922125 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6b4sb/crc-debug-7rthp"] Oct 04 10:01:50 crc kubenswrapper[4969]: I1004 10:01:50.924707 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:01:51 crc kubenswrapper[4969]: I1004 10:01:51.066725 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdqx2\" (UniqueName: \"kubernetes.io/projected/09844048-7fd2-4c20-8006-49a4fb31c6d8-kube-api-access-rdqx2\") pod \"crc-debug-7rthp\" (UID: \"09844048-7fd2-4c20-8006-49a4fb31c6d8\") " pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:01:51 crc kubenswrapper[4969]: I1004 10:01:51.066768 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09844048-7fd2-4c20-8006-49a4fb31c6d8-host\") pod \"crc-debug-7rthp\" (UID: \"09844048-7fd2-4c20-8006-49a4fb31c6d8\") " pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:01:51 crc kubenswrapper[4969]: I1004 10:01:51.168577 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdqx2\" (UniqueName: \"kubernetes.io/projected/09844048-7fd2-4c20-8006-49a4fb31c6d8-kube-api-access-rdqx2\") pod \"crc-debug-7rthp\" (UID: \"09844048-7fd2-4c20-8006-49a4fb31c6d8\") " pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:01:51 crc kubenswrapper[4969]: I1004 10:01:51.168629 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09844048-7fd2-4c20-8006-49a4fb31c6d8-host\") pod \"crc-debug-7rthp\" (UID: \"09844048-7fd2-4c20-8006-49a4fb31c6d8\") " pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:01:51 crc kubenswrapper[4969]: I1004 10:01:51.168771 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09844048-7fd2-4c20-8006-49a4fb31c6d8-host\") pod \"crc-debug-7rthp\" (UID: \"09844048-7fd2-4c20-8006-49a4fb31c6d8\") " pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:01:51 crc kubenswrapper[4969]: I1004 10:01:51.192417 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdqx2\" (UniqueName: \"kubernetes.io/projected/09844048-7fd2-4c20-8006-49a4fb31c6d8-kube-api-access-rdqx2\") pod \"crc-debug-7rthp\" (UID: \"09844048-7fd2-4c20-8006-49a4fb31c6d8\") " pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:01:51 crc kubenswrapper[4969]: I1004 10:01:51.254034 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:01:51 crc kubenswrapper[4969]: W1004 10:01:51.292375 4969 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09844048_7fd2_4c20_8006_49a4fb31c6d8.slice/crio-6a80a4176a29fceb2a1435ac059c1d96e5ebed84843141f1eb8e455b677779f4 WatchSource:0}: Error finding container 6a80a4176a29fceb2a1435ac059c1d96e5ebed84843141f1eb8e455b677779f4: Status 404 returned error can't find the container with id 6a80a4176a29fceb2a1435ac059c1d96e5ebed84843141f1eb8e455b677779f4 Oct 04 10:01:51 crc kubenswrapper[4969]: I1004 10:01:51.591706 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/crc-debug-7rthp" event={"ID":"09844048-7fd2-4c20-8006-49a4fb31c6d8","Type":"ContainerStarted","Data":"e493485269fc5f984c92b2cb126e270bdf3bad54200f121d9b0bb3b5ee8007ea"} Oct 04 10:01:51 crc kubenswrapper[4969]: I1004 10:01:51.591742 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/crc-debug-7rthp" event={"ID":"09844048-7fd2-4c20-8006-49a4fb31c6d8","Type":"ContainerStarted","Data":"6a80a4176a29fceb2a1435ac059c1d96e5ebed84843141f1eb8e455b677779f4"} Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.556343 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6b4sb/crc-debug-7rthp" podStartSLOduration=12.556323463 podStartE2EDuration="12.556323463s" podCreationTimestamp="2025-10-04 10:01:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:01:51.611700609 +0000 UTC m=+6339.365969423" watchObservedRunningTime="2025-10-04 10:02:02.556323463 +0000 UTC m=+6350.310592267" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.560030 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7p4kt"] Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.561919 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.584292 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p4kt"] Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.691689 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-catalog-content\") pod \"certified-operators-7p4kt\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.691827 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z45fj\" (UniqueName: \"kubernetes.io/projected/b820c199-abb8-43c5-965b-95c8560c1433-kube-api-access-z45fj\") pod \"certified-operators-7p4kt\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.691914 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-utilities\") pod \"certified-operators-7p4kt\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.793781 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-catalog-content\") pod \"certified-operators-7p4kt\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.793861 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z45fj\" (UniqueName: \"kubernetes.io/projected/b820c199-abb8-43c5-965b-95c8560c1433-kube-api-access-z45fj\") pod \"certified-operators-7p4kt\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.793914 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-utilities\") pod \"certified-operators-7p4kt\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.794482 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-utilities\") pod \"certified-operators-7p4kt\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.794482 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-catalog-content\") pod \"certified-operators-7p4kt\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.816061 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z45fj\" (UniqueName: \"kubernetes.io/projected/b820c199-abb8-43c5-965b-95c8560c1433-kube-api-access-z45fj\") pod \"certified-operators-7p4kt\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:02 crc kubenswrapper[4969]: I1004 10:02:02.881522 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:03 crc kubenswrapper[4969]: I1004 10:02:03.545743 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p4kt"] Oct 04 10:02:03 crc kubenswrapper[4969]: I1004 10:02:03.717232 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p4kt" event={"ID":"b820c199-abb8-43c5-965b-95c8560c1433","Type":"ContainerStarted","Data":"2b34dc06d1917c1e8946179c6ca0e7576cabe074b0e595064e9e15c1bf087f38"} Oct 04 10:02:04 crc kubenswrapper[4969]: I1004 10:02:04.729918 4969 generic.go:334] "Generic (PLEG): container finished" podID="b820c199-abb8-43c5-965b-95c8560c1433" containerID="c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f" exitCode=0 Oct 04 10:02:04 crc kubenswrapper[4969]: I1004 10:02:04.730098 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p4kt" event={"ID":"b820c199-abb8-43c5-965b-95c8560c1433","Type":"ContainerDied","Data":"c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f"} Oct 04 10:02:05 crc kubenswrapper[4969]: I1004 10:02:05.746326 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p4kt" event={"ID":"b820c199-abb8-43c5-965b-95c8560c1433","Type":"ContainerStarted","Data":"f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54"} Oct 04 10:02:06 crc kubenswrapper[4969]: I1004 10:02:06.769392 4969 generic.go:334] "Generic (PLEG): container finished" podID="b820c199-abb8-43c5-965b-95c8560c1433" containerID="f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54" exitCode=0 Oct 04 10:02:06 crc kubenswrapper[4969]: I1004 10:02:06.769667 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p4kt" event={"ID":"b820c199-abb8-43c5-965b-95c8560c1433","Type":"ContainerDied","Data":"f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54"} Oct 04 10:02:07 crc kubenswrapper[4969]: I1004 10:02:07.780768 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p4kt" event={"ID":"b820c199-abb8-43c5-965b-95c8560c1433","Type":"ContainerStarted","Data":"bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0"} Oct 04 10:02:07 crc kubenswrapper[4969]: I1004 10:02:07.806105 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7p4kt" podStartSLOduration=3.187887928 podStartE2EDuration="5.806090237s" podCreationTimestamp="2025-10-04 10:02:02 +0000 UTC" firstStartedPulling="2025-10-04 10:02:04.732342139 +0000 UTC m=+6352.486610943" lastFinishedPulling="2025-10-04 10:02:07.350544438 +0000 UTC m=+6355.104813252" observedRunningTime="2025-10-04 10:02:07.800228204 +0000 UTC m=+6355.554497018" watchObservedRunningTime="2025-10-04 10:02:07.806090237 +0000 UTC m=+6355.560359041" Oct 04 10:02:12 crc kubenswrapper[4969]: I1004 10:02:12.882756 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:12 crc kubenswrapper[4969]: I1004 10:02:12.884459 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:12 crc kubenswrapper[4969]: I1004 10:02:12.936549 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:13 crc kubenswrapper[4969]: I1004 10:02:13.878295 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:13 crc kubenswrapper[4969]: I1004 10:02:13.926089 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7p4kt"] Oct 04 10:02:15 crc kubenswrapper[4969]: I1004 10:02:15.842469 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7p4kt" podUID="b820c199-abb8-43c5-965b-95c8560c1433" containerName="registry-server" containerID="cri-o://bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0" gracePeriod=2 Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.455803 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.606966 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-utilities\") pod \"b820c199-abb8-43c5-965b-95c8560c1433\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.607026 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-catalog-content\") pod \"b820c199-abb8-43c5-965b-95c8560c1433\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.607273 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z45fj\" (UniqueName: \"kubernetes.io/projected/b820c199-abb8-43c5-965b-95c8560c1433-kube-api-access-z45fj\") pod \"b820c199-abb8-43c5-965b-95c8560c1433\" (UID: \"b820c199-abb8-43c5-965b-95c8560c1433\") " Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.608455 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-utilities" (OuterVolumeSpecName: "utilities") pod "b820c199-abb8-43c5-965b-95c8560c1433" (UID: "b820c199-abb8-43c5-965b-95c8560c1433"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.612884 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b820c199-abb8-43c5-965b-95c8560c1433-kube-api-access-z45fj" (OuterVolumeSpecName: "kube-api-access-z45fj") pod "b820c199-abb8-43c5-965b-95c8560c1433" (UID: "b820c199-abb8-43c5-965b-95c8560c1433"). InnerVolumeSpecName "kube-api-access-z45fj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.660933 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b820c199-abb8-43c5-965b-95c8560c1433" (UID: "b820c199-abb8-43c5-965b-95c8560c1433"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.709271 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z45fj\" (UniqueName: \"kubernetes.io/projected/b820c199-abb8-43c5-965b-95c8560c1433-kube-api-access-z45fj\") on node \"crc\" DevicePath \"\"" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.709306 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.709318 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b820c199-abb8-43c5-965b-95c8560c1433-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.854173 4969 generic.go:334] "Generic (PLEG): container finished" podID="b820c199-abb8-43c5-965b-95c8560c1433" containerID="bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0" exitCode=0 Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.854217 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p4kt" event={"ID":"b820c199-abb8-43c5-965b-95c8560c1433","Type":"ContainerDied","Data":"bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0"} Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.854245 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p4kt" event={"ID":"b820c199-abb8-43c5-965b-95c8560c1433","Type":"ContainerDied","Data":"2b34dc06d1917c1e8946179c6ca0e7576cabe074b0e595064e9e15c1bf087f38"} Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.854263 4969 scope.go:117] "RemoveContainer" containerID="bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.854387 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p4kt" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.901961 4969 scope.go:117] "RemoveContainer" containerID="f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.914443 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7p4kt"] Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.923560 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7p4kt"] Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.961349 4969 scope.go:117] "RemoveContainer" containerID="c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.993601 4969 scope.go:117] "RemoveContainer" containerID="bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0" Oct 04 10:02:16 crc kubenswrapper[4969]: E1004 10:02:16.995250 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0\": container with ID starting with bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0 not found: ID does not exist" containerID="bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.995281 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0"} err="failed to get container status \"bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0\": rpc error: code = NotFound desc = could not find container \"bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0\": container with ID starting with bdfae0a7f018e0996b8f060b2f6259f63a8186806e162343dd0da0166fa7bfb0 not found: ID does not exist" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.995311 4969 scope.go:117] "RemoveContainer" containerID="f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54" Oct 04 10:02:16 crc kubenswrapper[4969]: E1004 10:02:16.995677 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54\": container with ID starting with f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54 not found: ID does not exist" containerID="f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.995697 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54"} err="failed to get container status \"f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54\": rpc error: code = NotFound desc = could not find container \"f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54\": container with ID starting with f3b2c6ea155db2c02dee5dbcad83a6d5f5fbb64e9bb8109fb3c5b020544c4a54 not found: ID does not exist" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.995711 4969 scope.go:117] "RemoveContainer" containerID="c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f" Oct 04 10:02:16 crc kubenswrapper[4969]: E1004 10:02:16.996059 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f\": container with ID starting with c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f not found: ID does not exist" containerID="c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f" Oct 04 10:02:16 crc kubenswrapper[4969]: I1004 10:02:16.996081 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f"} err="failed to get container status \"c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f\": rpc error: code = NotFound desc = could not find container \"c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f\": container with ID starting with c90b5d268d0395e6ce1b3c67c51bda696ec9599277149d6a91dec0f80042aa6f not found: ID does not exist" Oct 04 10:02:17 crc kubenswrapper[4969]: I1004 10:02:17.074835 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b820c199-abb8-43c5-965b-95c8560c1433" path="/var/lib/kubelet/pods/b820c199-abb8-43c5-965b-95c8560c1433/volumes" Oct 04 10:02:19 crc kubenswrapper[4969]: I1004 10:02:19.666471 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:02:19 crc kubenswrapper[4969]: I1004 10:02:19.666919 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:02:49 crc kubenswrapper[4969]: I1004 10:02:49.666274 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:02:49 crc kubenswrapper[4969]: I1004 10:02:49.666940 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:03:07 crc kubenswrapper[4969]: I1004 10:03:07.024460 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dbff44956-qlvk2_d46c925d-1b2f-4696-b099-f3a549dda18b/barbican-api/0.log" Oct 04 10:03:07 crc kubenswrapper[4969]: I1004 10:03:07.168940 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dbff44956-qlvk2_d46c925d-1b2f-4696-b099-f3a549dda18b/barbican-api-log/0.log" Oct 04 10:03:07 crc kubenswrapper[4969]: I1004 10:03:07.236403 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7d5b5f7c58-kgs6c_f2e7b743-71f9-46c8-aeb7-450c63bf22ca/barbican-keystone-listener/0.log" Oct 04 10:03:07 crc kubenswrapper[4969]: I1004 10:03:07.439261 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7d5b5f7c58-kgs6c_f2e7b743-71f9-46c8-aeb7-450c63bf22ca/barbican-keystone-listener-log/0.log" Oct 04 10:03:07 crc kubenswrapper[4969]: I1004 10:03:07.448811 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9fb8f975c-4l4lz_3f74b083-ba83-4f80-a8a6-e5e10e6dde28/barbican-worker/0.log" Oct 04 10:03:07 crc kubenswrapper[4969]: I1004 10:03:07.615733 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9fb8f975c-4l4lz_3f74b083-ba83-4f80-a8a6-e5e10e6dde28/barbican-worker-log/0.log" Oct 04 10:03:07 crc kubenswrapper[4969]: I1004 10:03:07.688820 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-q68gw_6e5dcf0e-60ad-4ddf-b0a7-e504f4566ce7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:07 crc kubenswrapper[4969]: I1004 10:03:07.921856 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/ceilometer-central-agent/1.log" Oct 04 10:03:07 crc kubenswrapper[4969]: I1004 10:03:07.941470 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/ceilometer-central-agent/0.log" Oct 04 10:03:08 crc kubenswrapper[4969]: I1004 10:03:08.115395 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/sg-core/0.log" Oct 04 10:03:08 crc kubenswrapper[4969]: I1004 10:03:08.124903 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/ceilometer-notification-agent/0.log" Oct 04 10:03:08 crc kubenswrapper[4969]: I1004 10:03:08.174000 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a525e7b-7a1c-4671-a6cc-e81bd316fd48/proxy-httpd/0.log" Oct 04 10:03:08 crc kubenswrapper[4969]: I1004 10:03:08.398621 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_755c8160-b3df-4ba0-8d91-46d4af514dfc/cinder-api-log/0.log" Oct 04 10:03:08 crc kubenswrapper[4969]: I1004 10:03:08.695190 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_755c8160-b3df-4ba0-8d91-46d4af514dfc/cinder-api/0.log" Oct 04 10:03:08 crc kubenswrapper[4969]: I1004 10:03:08.702142 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a7183f88-c473-4e74-8010-f8e7a380b35f/cinder-scheduler/0.log" Oct 04 10:03:08 crc kubenswrapper[4969]: I1004 10:03:08.943608 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a7183f88-c473-4e74-8010-f8e7a380b35f/probe/0.log" Oct 04 10:03:08 crc kubenswrapper[4969]: I1004 10:03:08.947555 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-fxzl2_48e0d7b9-c122-4cee-84b3-1e89a72af2ad/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:09 crc kubenswrapper[4969]: I1004 10:03:09.177277 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2d2sk_85fdbfa3-25a5-4a92-9473-198d3cf416c8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:09 crc kubenswrapper[4969]: I1004 10:03:09.333657 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-t796n_19b4ae96-fa76-4d96-b4e0-d0058345fc23/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:09 crc kubenswrapper[4969]: I1004 10:03:09.447147 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bd8b95f47-sjpqk_342bf40f-63d2-47cd-82c8-5c2aec79d3d2/init/0.log" Oct 04 10:03:09 crc kubenswrapper[4969]: I1004 10:03:09.635794 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bd8b95f47-sjpqk_342bf40f-63d2-47cd-82c8-5c2aec79d3d2/init/0.log" Oct 04 10:03:09 crc kubenswrapper[4969]: I1004 10:03:09.840159 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bd8b95f47-sjpqk_342bf40f-63d2-47cd-82c8-5c2aec79d3d2/dnsmasq-dns/0.log" Oct 04 10:03:09 crc kubenswrapper[4969]: I1004 10:03:09.851210 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-4r4vz_a71c86d5-acae-401d-bea7-8cee0956d990/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:10 crc kubenswrapper[4969]: I1004 10:03:10.058408 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7c627df9-6989-44d8-b72f-128992d65bd9/glance-httpd/0.log" Oct 04 10:03:10 crc kubenswrapper[4969]: I1004 10:03:10.064837 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7c627df9-6989-44d8-b72f-128992d65bd9/glance-log/0.log" Oct 04 10:03:10 crc kubenswrapper[4969]: I1004 10:03:10.259749 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7ebb3f7a-584d-4c2d-a26f-f66481883b6e/glance-httpd/0.log" Oct 04 10:03:10 crc kubenswrapper[4969]: I1004 10:03:10.303191 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7ebb3f7a-584d-4c2d-a26f-f66481883b6e/glance-log/0.log" Oct 04 10:03:10 crc kubenswrapper[4969]: I1004 10:03:10.449344 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f94787ff8-mkj4z_3344b512-4dea-4112-a8b1-7fa337ef81ae/horizon/0.log" Oct 04 10:03:10 crc kubenswrapper[4969]: I1004 10:03:10.551008 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-9kbsw_a7f1a5ed-9324-4eb1-b561-d65fca042f4d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:10 crc kubenswrapper[4969]: I1004 10:03:10.758980 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-cqqr6_579fed39-17e4-412c-b76b-139db1b5938e/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:10 crc kubenswrapper[4969]: I1004 10:03:10.962992 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29326141-knk55_cbbc9d5a-0ea4-4761-b48f-79aaeea49a6a/keystone-cron/0.log" Oct 04 10:03:11 crc kubenswrapper[4969]: I1004 10:03:11.345194 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f94787ff8-mkj4z_3344b512-4dea-4112-a8b1-7fa337ef81ae/horizon-log/0.log" Oct 04 10:03:11 crc kubenswrapper[4969]: I1004 10:03:11.378003 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29326201-l86vc_b7d3f669-1fd4-4a36-90fe-0bf25c7aa9ad/keystone-cron/0.log" Oct 04 10:03:11 crc kubenswrapper[4969]: I1004 10:03:11.533139 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-699ff78d6b-dv46q_7327643b-aa39-4085-88c4-bc4411b8d832/keystone-api/0.log" Oct 04 10:03:11 crc kubenswrapper[4969]: I1004 10:03:11.618648 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_5385fdd1-eb06-4d69-a383-236a55505108/kube-state-metrics/0.log" Oct 04 10:03:11 crc kubenswrapper[4969]: I1004 10:03:11.705363 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-r2ss9_07ace2b0-20d8-4b9a-8f93-4649a3eee5bb/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:12 crc kubenswrapper[4969]: I1004 10:03:12.271845 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-78c8455845-gfrwq_e82aaa27-6e5c-43d8-9ed2-c0958de485c7/neutron-httpd/0.log" Oct 04 10:03:12 crc kubenswrapper[4969]: I1004 10:03:12.285462 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-78c8455845-gfrwq_e82aaa27-6e5c-43d8-9ed2-c0958de485c7/neutron-api/0.log" Oct 04 10:03:12 crc kubenswrapper[4969]: I1004 10:03:12.296862 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-whwq4_87bcc799-a7f6-4d1f-83d5-53b3ada4e8a6/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:13 crc kubenswrapper[4969]: I1004 10:03:13.262352 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_60f12325-d22b-4986-a653-854d290f0adc/nova-cell0-conductor-conductor/0.log" Oct 04 10:03:13 crc kubenswrapper[4969]: I1004 10:03:13.958209 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ac48f46b-0304-424d-b394-a1e2a274230a/nova-cell1-conductor-conductor/0.log" Oct 04 10:03:14 crc kubenswrapper[4969]: I1004 10:03:14.164411 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e5b33466-d882-4113-8c9a-bc1047ee6b9a/nova-api-log/0.log" Oct 04 10:03:14 crc kubenswrapper[4969]: I1004 10:03:14.521604 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_2075dd48-ec13-4ebb-841b-3a2d82724f98/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 10:03:14 crc kubenswrapper[4969]: I1004 10:03:14.660903 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e5b33466-d882-4113-8c9a-bc1047ee6b9a/nova-api-api/0.log" Oct 04 10:03:14 crc kubenswrapper[4969]: I1004 10:03:14.667077 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-p9v4f_4ce3e260-2be8-435b-8a24-7d06614c0844/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:14 crc kubenswrapper[4969]: I1004 10:03:14.971666 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b7574220-81ab-4bde-b17b-c15d3339bfd6/nova-metadata-log/0.log" Oct 04 10:03:15 crc kubenswrapper[4969]: I1004 10:03:15.459126 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_82fe815c-3f06-499f-9e3c-5a3b2f00d932/mysql-bootstrap/0.log" Oct 04 10:03:15 crc kubenswrapper[4969]: I1004 10:03:15.550339 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_9ef1ee4f-398a-49d8-a569-2d5bcc4cf073/nova-scheduler-scheduler/0.log" Oct 04 10:03:15 crc kubenswrapper[4969]: I1004 10:03:15.614067 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_82fe815c-3f06-499f-9e3c-5a3b2f00d932/mysql-bootstrap/0.log" Oct 04 10:03:15 crc kubenswrapper[4969]: I1004 10:03:15.720727 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_82fe815c-3f06-499f-9e3c-5a3b2f00d932/galera/0.log" Oct 04 10:03:15 crc kubenswrapper[4969]: I1004 10:03:15.895966 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8a3d9463-8792-487d-b4a5-1b61ab3310f6/mysql-bootstrap/0.log" Oct 04 10:03:16 crc kubenswrapper[4969]: I1004 10:03:16.135108 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8a3d9463-8792-487d-b4a5-1b61ab3310f6/galera/0.log" Oct 04 10:03:16 crc kubenswrapper[4969]: I1004 10:03:16.152876 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8a3d9463-8792-487d-b4a5-1b61ab3310f6/mysql-bootstrap/0.log" Oct 04 10:03:16 crc kubenswrapper[4969]: I1004 10:03:16.337224 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f5532f51-e305-4741-8661-25029faf98f2/openstackclient/0.log" Oct 04 10:03:16 crc kubenswrapper[4969]: I1004 10:03:16.560436 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-fms7j_452b4c0a-fcb1-48ed-8c3a-87c34904a0a5/ovn-controller/0.log" Oct 04 10:03:16 crc kubenswrapper[4969]: I1004 10:03:16.796276 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-zssjt_0b09ae48-946d-458d-a119-194f28bd1081/openstack-network-exporter/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.013008 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p9gfr_3f6d0bb5-43f9-4c31-838b-f71927d68b08/ovsdb-server-init/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.120318 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_41d6e66e-6079-4278-a26c-04a299dd6d9d/memcached/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.173183 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p9gfr_3f6d0bb5-43f9-4c31-838b-f71927d68b08/ovsdb-server-init/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.303134 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p9gfr_3f6d0bb5-43f9-4c31-838b-f71927d68b08/ovsdb-server/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.353908 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b7574220-81ab-4bde-b17b-c15d3339bfd6/nova-metadata-metadata/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.536124 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4hrwl_ceaa1102-1d9c-40ec-a6a7-b31ca499c55e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.546769 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p9gfr_3f6d0bb5-43f9-4c31-838b-f71927d68b08/ovs-vswitchd/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.656583 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_138cab2b-aabb-4736-a566-d9aee97753c0/openstack-network-exporter/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.714888 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_138cab2b-aabb-4736-a566-d9aee97753c0/ovn-northd/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.724865 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8d2f95cc-1c69-4731-90a9-a57b693b87ac/openstack-network-exporter/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.835321 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8d2f95cc-1c69-4731-90a9-a57b693b87ac/ovsdbserver-nb/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.904855 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d179b5df-13f6-4e3f-bbd5-d210ba6b2f88/openstack-network-exporter/0.log" Oct 04 10:03:17 crc kubenswrapper[4969]: I1004 10:03:17.930210 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d179b5df-13f6-4e3f-bbd5-d210ba6b2f88/ovsdbserver-sb/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.307715 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/init-config-reloader/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.345021 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-569b746d9d-r7c45_d6425541-fb00-4545-a4df-64aba5d7e5ab/placement-api/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.352569 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-569b746d9d-r7c45_d6425541-fb00-4545-a4df-64aba5d7e5ab/placement-log/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.492952 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/init-config-reloader/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.500292 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/config-reloader/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.510697 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/prometheus/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.536904 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9c93fd81-fe7a-4e9f-83ef-166655d4d8d0/thanos-sidecar/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.660198 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1fdac808-4983-49ff-9975-ebc0ad82bb27/setup-container/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.814097 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1fdac808-4983-49ff-9975-ebc0ad82bb27/setup-container/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.853084 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1fdac808-4983-49ff-9975-ebc0ad82bb27/rabbitmq/0.log" Oct 04 10:03:18 crc kubenswrapper[4969]: I1004 10:03:18.919915 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_28264a08-8168-419d-a20d-8f94a5c4ed77/setup-container/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.082164 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_28264a08-8168-419d-a20d-8f94a5c4ed77/setup-container/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.089242 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_28264a08-8168-419d-a20d-8f94a5c4ed77/rabbitmq/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.136451 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7379ac07-e6f6-4218-85c2-76883f50cf0e/setup-container/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.280596 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7379ac07-e6f6-4218-85c2-76883f50cf0e/setup-container/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.291668 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7379ac07-e6f6-4218-85c2-76883f50cf0e/rabbitmq/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.343529 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-2gtnm_942f1580-e8dc-4f73-aa37-e5a2cde068c6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.475259 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-j9fsg_562cc378-72b8-42ea-9822-4a5aa6cebbb6/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.611388 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-qqp4z_a4b37224-1961-43f0-bbc9-efbc635bc063/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.666279 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.666342 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.666387 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.667174 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8fdf0e4cb48a4d351d146bddccf73a2a4695a141ecf2ec805971c3bfed433381"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.667230 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://8fdf0e4cb48a4d351d146bddccf73a2a4695a141ecf2ec805971c3bfed433381" gracePeriod=600 Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.716498 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kclpg_57aac488-900c-47d7-abc5-0349f1004655/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:19 crc kubenswrapper[4969]: I1004 10:03:19.804614 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dhdzp_b0fface3-5e9a-4a86-9863-8c7ad3c21151/ssh-known-hosts-edpm-deployment/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:19.999898 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-75f87ddd45-2md8q_f2887852-0cd7-476e-8c4e-1ed98c66ede4/proxy-server/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.148797 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-75f87ddd45-2md8q_f2887852-0cd7-476e-8c4e-1ed98c66ede4/proxy-httpd/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.246335 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-b822c_67329c10-3dda-4a99-a964-cc8039752729/swift-ring-rebalance/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.372520 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/account-reaper/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.375161 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/account-auditor/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.505643 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/account-server/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.514410 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="8fdf0e4cb48a4d351d146bddccf73a2a4695a141ecf2ec805971c3bfed433381" exitCode=0 Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.514458 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"8fdf0e4cb48a4d351d146bddccf73a2a4695a141ecf2ec805971c3bfed433381"} Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.514506 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerStarted","Data":"779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b"} Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.514526 4969 scope.go:117] "RemoveContainer" containerID="0753955b7a631a399016582684d13eb862e4d71c7bead9fcaee01829e1dae7c8" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.516944 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/account-replicator/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.556910 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/container-auditor/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.648307 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/container-replicator/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.682334 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/container-server/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.745399 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/container-updater/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.782541 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-auditor/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.831050 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-expirer/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.899743 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-replicator/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.904071 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-server/0.log" Oct 04 10:03:20 crc kubenswrapper[4969]: I1004 10:03:20.956193 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/object-updater/0.log" Oct 04 10:03:21 crc kubenswrapper[4969]: I1004 10:03:21.022634 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/rsync/0.log" Oct 04 10:03:21 crc kubenswrapper[4969]: I1004 10:03:21.117099 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c7dc30e8-fc6b-4fb3-95b4-df2f1da176a4/swift-recon-cron/0.log" Oct 04 10:03:21 crc kubenswrapper[4969]: I1004 10:03:21.160011 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-k45fp_eefed526-feaa-4158-94be-19220b2801e2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:21 crc kubenswrapper[4969]: I1004 10:03:21.314386 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_6348f782-6e3b-4c67-b252-2276628c14d6/tempest-tests-tempest-tests-runner/0.log" Oct 04 10:03:21 crc kubenswrapper[4969]: I1004 10:03:21.379744 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_9e320e23-ff99-4d9f-b2aa-25c7d2420b88/test-operator-logs-container/0.log" Oct 04 10:03:21 crc kubenswrapper[4969]: I1004 10:03:21.546865 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-qjjb7_8c7813c0-2dcd-4d30-9f8c-fac53077417f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 10:03:22 crc kubenswrapper[4969]: I1004 10:03:22.440717 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_e7a2121b-d2c2-4735-adb7-9b2bfb7a1fac/watcher-applier/0.log" Oct 04 10:03:22 crc kubenswrapper[4969]: I1004 10:03:22.824900 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_b36dc326-a420-42e1-a274-9e82641ee148/watcher-api-log/0.log" Oct 04 10:03:24 crc kubenswrapper[4969]: I1004 10:03:24.990861 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_6f21dc45-2a75-427d-a2c1-632c0fa0428c/watcher-decision-engine/0.log" Oct 04 10:03:25 crc kubenswrapper[4969]: I1004 10:03:25.955569 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_b36dc326-a420-42e1-a274-9e82641ee148/watcher-api/0.log" Oct 04 10:03:35 crc kubenswrapper[4969]: I1004 10:03:35.368294 4969 scope.go:117] "RemoveContainer" containerID="863c54823502497d8c50f93de4e3fdd24366712e00070fd56ded1e109581bdee" Oct 04 10:03:50 crc kubenswrapper[4969]: I1004 10:03:50.796657 4969 generic.go:334] "Generic (PLEG): container finished" podID="09844048-7fd2-4c20-8006-49a4fb31c6d8" containerID="e493485269fc5f984c92b2cb126e270bdf3bad54200f121d9b0bb3b5ee8007ea" exitCode=0 Oct 04 10:03:50 crc kubenswrapper[4969]: I1004 10:03:50.796731 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/crc-debug-7rthp" event={"ID":"09844048-7fd2-4c20-8006-49a4fb31c6d8","Type":"ContainerDied","Data":"e493485269fc5f984c92b2cb126e270bdf3bad54200f121d9b0bb3b5ee8007ea"} Oct 04 10:03:51 crc kubenswrapper[4969]: I1004 10:03:51.942645 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:03:51 crc kubenswrapper[4969]: I1004 10:03:51.987056 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6b4sb/crc-debug-7rthp"] Oct 04 10:03:51 crc kubenswrapper[4969]: I1004 10:03:51.994408 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6b4sb/crc-debug-7rthp"] Oct 04 10:03:52 crc kubenswrapper[4969]: I1004 10:03:52.077954 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdqx2\" (UniqueName: \"kubernetes.io/projected/09844048-7fd2-4c20-8006-49a4fb31c6d8-kube-api-access-rdqx2\") pod \"09844048-7fd2-4c20-8006-49a4fb31c6d8\" (UID: \"09844048-7fd2-4c20-8006-49a4fb31c6d8\") " Oct 04 10:03:52 crc kubenswrapper[4969]: I1004 10:03:52.078032 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09844048-7fd2-4c20-8006-49a4fb31c6d8-host\") pod \"09844048-7fd2-4c20-8006-49a4fb31c6d8\" (UID: \"09844048-7fd2-4c20-8006-49a4fb31c6d8\") " Oct 04 10:03:52 crc kubenswrapper[4969]: I1004 10:03:52.078181 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09844048-7fd2-4c20-8006-49a4fb31c6d8-host" (OuterVolumeSpecName: "host") pod "09844048-7fd2-4c20-8006-49a4fb31c6d8" (UID: "09844048-7fd2-4c20-8006-49a4fb31c6d8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:03:52 crc kubenswrapper[4969]: I1004 10:03:52.078697 4969 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09844048-7fd2-4c20-8006-49a4fb31c6d8-host\") on node \"crc\" DevicePath \"\"" Oct 04 10:03:52 crc kubenswrapper[4969]: I1004 10:03:52.087755 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09844048-7fd2-4c20-8006-49a4fb31c6d8-kube-api-access-rdqx2" (OuterVolumeSpecName: "kube-api-access-rdqx2") pod "09844048-7fd2-4c20-8006-49a4fb31c6d8" (UID: "09844048-7fd2-4c20-8006-49a4fb31c6d8"). InnerVolumeSpecName "kube-api-access-rdqx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:03:52 crc kubenswrapper[4969]: I1004 10:03:52.181033 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdqx2\" (UniqueName: \"kubernetes.io/projected/09844048-7fd2-4c20-8006-49a4fb31c6d8-kube-api-access-rdqx2\") on node \"crc\" DevicePath \"\"" Oct 04 10:03:52 crc kubenswrapper[4969]: I1004 10:03:52.825588 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a80a4176a29fceb2a1435ac059c1d96e5ebed84843141f1eb8e455b677779f4" Oct 04 10:03:52 crc kubenswrapper[4969]: I1004 10:03:52.825699 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-7rthp" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.074876 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09844048-7fd2-4c20-8006-49a4fb31c6d8" path="/var/lib/kubelet/pods/09844048-7fd2-4c20-8006-49a4fb31c6d8/volumes" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.214225 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6b4sb/crc-debug-tjxnw"] Oct 04 10:03:53 crc kubenswrapper[4969]: E1004 10:03:53.214761 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09844048-7fd2-4c20-8006-49a4fb31c6d8" containerName="container-00" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.214786 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="09844048-7fd2-4c20-8006-49a4fb31c6d8" containerName="container-00" Oct 04 10:03:53 crc kubenswrapper[4969]: E1004 10:03:53.214804 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b820c199-abb8-43c5-965b-95c8560c1433" containerName="extract-content" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.214813 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b820c199-abb8-43c5-965b-95c8560c1433" containerName="extract-content" Oct 04 10:03:53 crc kubenswrapper[4969]: E1004 10:03:53.214826 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b820c199-abb8-43c5-965b-95c8560c1433" containerName="extract-utilities" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.214835 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b820c199-abb8-43c5-965b-95c8560c1433" containerName="extract-utilities" Oct 04 10:03:53 crc kubenswrapper[4969]: E1004 10:03:53.214850 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b820c199-abb8-43c5-965b-95c8560c1433" containerName="registry-server" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.214860 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="b820c199-abb8-43c5-965b-95c8560c1433" containerName="registry-server" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.215117 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="09844048-7fd2-4c20-8006-49a4fb31c6d8" containerName="container-00" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.215144 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="b820c199-abb8-43c5-965b-95c8560c1433" containerName="registry-server" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.215956 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.304969 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-host\") pod \"crc-debug-tjxnw\" (UID: \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\") " pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.305032 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzrwz\" (UniqueName: \"kubernetes.io/projected/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-kube-api-access-rzrwz\") pod \"crc-debug-tjxnw\" (UID: \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\") " pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.407416 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-host\") pod \"crc-debug-tjxnw\" (UID: \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\") " pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.407531 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzrwz\" (UniqueName: \"kubernetes.io/projected/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-kube-api-access-rzrwz\") pod \"crc-debug-tjxnw\" (UID: \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\") " pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.407647 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-host\") pod \"crc-debug-tjxnw\" (UID: \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\") " pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.431104 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzrwz\" (UniqueName: \"kubernetes.io/projected/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-kube-api-access-rzrwz\") pod \"crc-debug-tjxnw\" (UID: \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\") " pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.537472 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.839003 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" event={"ID":"9c641277-250b-4eed-a7ce-eaa4e08a7cf8","Type":"ContainerStarted","Data":"569221082ddcbb8428342418c87475833e6e8eecb2dcfcae022492c71f635d4a"} Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.839335 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" event={"ID":"9c641277-250b-4eed-a7ce-eaa4e08a7cf8","Type":"ContainerStarted","Data":"9152fb95ddb2f85a5916036534e66cf1b940481d3fa97140f4790bc2c04acbcc"} Oct 04 10:03:53 crc kubenswrapper[4969]: I1004 10:03:53.858165 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" podStartSLOduration=0.858131512 podStartE2EDuration="858.131512ms" podCreationTimestamp="2025-10-04 10:03:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:03:53.854189825 +0000 UTC m=+6461.608458649" watchObservedRunningTime="2025-10-04 10:03:53.858131512 +0000 UTC m=+6461.612400366" Oct 04 10:03:54 crc kubenswrapper[4969]: I1004 10:03:54.846984 4969 generic.go:334] "Generic (PLEG): container finished" podID="9c641277-250b-4eed-a7ce-eaa4e08a7cf8" containerID="569221082ddcbb8428342418c87475833e6e8eecb2dcfcae022492c71f635d4a" exitCode=0 Oct 04 10:03:54 crc kubenswrapper[4969]: I1004 10:03:54.847062 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" event={"ID":"9c641277-250b-4eed-a7ce-eaa4e08a7cf8","Type":"ContainerDied","Data":"569221082ddcbb8428342418c87475833e6e8eecb2dcfcae022492c71f635d4a"} Oct 04 10:03:55 crc kubenswrapper[4969]: I1004 10:03:55.957635 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:03:56 crc kubenswrapper[4969]: I1004 10:03:56.056133 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzrwz\" (UniqueName: \"kubernetes.io/projected/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-kube-api-access-rzrwz\") pod \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\" (UID: \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\") " Oct 04 10:03:56 crc kubenswrapper[4969]: I1004 10:03:56.056381 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-host\") pod \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\" (UID: \"9c641277-250b-4eed-a7ce-eaa4e08a7cf8\") " Oct 04 10:03:56 crc kubenswrapper[4969]: I1004 10:03:56.056805 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-host" (OuterVolumeSpecName: "host") pod "9c641277-250b-4eed-a7ce-eaa4e08a7cf8" (UID: "9c641277-250b-4eed-a7ce-eaa4e08a7cf8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:03:56 crc kubenswrapper[4969]: I1004 10:03:56.063613 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-kube-api-access-rzrwz" (OuterVolumeSpecName: "kube-api-access-rzrwz") pod "9c641277-250b-4eed-a7ce-eaa4e08a7cf8" (UID: "9c641277-250b-4eed-a7ce-eaa4e08a7cf8"). InnerVolumeSpecName "kube-api-access-rzrwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:03:56 crc kubenswrapper[4969]: I1004 10:03:56.158884 4969 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-host\") on node \"crc\" DevicePath \"\"" Oct 04 10:03:56 crc kubenswrapper[4969]: I1004 10:03:56.159151 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzrwz\" (UniqueName: \"kubernetes.io/projected/9c641277-250b-4eed-a7ce-eaa4e08a7cf8-kube-api-access-rzrwz\") on node \"crc\" DevicePath \"\"" Oct 04 10:03:56 crc kubenswrapper[4969]: I1004 10:03:56.875496 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" event={"ID":"9c641277-250b-4eed-a7ce-eaa4e08a7cf8","Type":"ContainerDied","Data":"9152fb95ddb2f85a5916036534e66cf1b940481d3fa97140f4790bc2c04acbcc"} Oct 04 10:03:56 crc kubenswrapper[4969]: I1004 10:03:56.875542 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9152fb95ddb2f85a5916036534e66cf1b940481d3fa97140f4790bc2c04acbcc" Oct 04 10:03:56 crc kubenswrapper[4969]: I1004 10:03:56.875606 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-tjxnw" Oct 04 10:04:02 crc kubenswrapper[4969]: I1004 10:04:02.690554 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6b4sb/crc-debug-tjxnw"] Oct 04 10:04:02 crc kubenswrapper[4969]: I1004 10:04:02.697720 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6b4sb/crc-debug-tjxnw"] Oct 04 10:04:03 crc kubenswrapper[4969]: I1004 10:04:03.069339 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c641277-250b-4eed-a7ce-eaa4e08a7cf8" path="/var/lib/kubelet/pods/9c641277-250b-4eed-a7ce-eaa4e08a7cf8/volumes" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.012749 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6b4sb/crc-debug-q4tll"] Oct 04 10:04:04 crc kubenswrapper[4969]: E1004 10:04:04.013170 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c641277-250b-4eed-a7ce-eaa4e08a7cf8" containerName="container-00" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.013184 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c641277-250b-4eed-a7ce-eaa4e08a7cf8" containerName="container-00" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.013409 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c641277-250b-4eed-a7ce-eaa4e08a7cf8" containerName="container-00" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.014591 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.104441 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdxjg\" (UniqueName: \"kubernetes.io/projected/feec8c4b-e98c-4fdd-bf70-235b413d7e90-kube-api-access-pdxjg\") pod \"crc-debug-q4tll\" (UID: \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\") " pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.104541 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feec8c4b-e98c-4fdd-bf70-235b413d7e90-host\") pod \"crc-debug-q4tll\" (UID: \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\") " pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.206659 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdxjg\" (UniqueName: \"kubernetes.io/projected/feec8c4b-e98c-4fdd-bf70-235b413d7e90-kube-api-access-pdxjg\") pod \"crc-debug-q4tll\" (UID: \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\") " pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.206785 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feec8c4b-e98c-4fdd-bf70-235b413d7e90-host\") pod \"crc-debug-q4tll\" (UID: \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\") " pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.208018 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feec8c4b-e98c-4fdd-bf70-235b413d7e90-host\") pod \"crc-debug-q4tll\" (UID: \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\") " pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.241795 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdxjg\" (UniqueName: \"kubernetes.io/projected/feec8c4b-e98c-4fdd-bf70-235b413d7e90-kube-api-access-pdxjg\") pod \"crc-debug-q4tll\" (UID: \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\") " pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.330809 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.955219 4969 generic.go:334] "Generic (PLEG): container finished" podID="feec8c4b-e98c-4fdd-bf70-235b413d7e90" containerID="37b3249dedfa6f019e03735c72c5d26a9bd04c4863f8a1d657f1bb40bb317221" exitCode=0 Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.955319 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/crc-debug-q4tll" event={"ID":"feec8c4b-e98c-4fdd-bf70-235b413d7e90","Type":"ContainerDied","Data":"37b3249dedfa6f019e03735c72c5d26a9bd04c4863f8a1d657f1bb40bb317221"} Oct 04 10:04:04 crc kubenswrapper[4969]: I1004 10:04:04.955757 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/crc-debug-q4tll" event={"ID":"feec8c4b-e98c-4fdd-bf70-235b413d7e90","Type":"ContainerStarted","Data":"293979e477af11c7a914ce81fcfec1c1dd305cd7b26e4934ff14c6422461b9d2"} Oct 04 10:04:05 crc kubenswrapper[4969]: I1004 10:04:05.012869 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6b4sb/crc-debug-q4tll"] Oct 04 10:04:05 crc kubenswrapper[4969]: I1004 10:04:05.023672 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6b4sb/crc-debug-q4tll"] Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.084733 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.245382 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feec8c4b-e98c-4fdd-bf70-235b413d7e90-host\") pod \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\" (UID: \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\") " Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.245531 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdxjg\" (UniqueName: \"kubernetes.io/projected/feec8c4b-e98c-4fdd-bf70-235b413d7e90-kube-api-access-pdxjg\") pod \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\" (UID: \"feec8c4b-e98c-4fdd-bf70-235b413d7e90\") " Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.245538 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/feec8c4b-e98c-4fdd-bf70-235b413d7e90-host" (OuterVolumeSpecName: "host") pod "feec8c4b-e98c-4fdd-bf70-235b413d7e90" (UID: "feec8c4b-e98c-4fdd-bf70-235b413d7e90"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.246595 4969 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feec8c4b-e98c-4fdd-bf70-235b413d7e90-host\") on node \"crc\" DevicePath \"\"" Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.250557 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feec8c4b-e98c-4fdd-bf70-235b413d7e90-kube-api-access-pdxjg" (OuterVolumeSpecName: "kube-api-access-pdxjg") pod "feec8c4b-e98c-4fdd-bf70-235b413d7e90" (UID: "feec8c4b-e98c-4fdd-bf70-235b413d7e90"). InnerVolumeSpecName "kube-api-access-pdxjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.348287 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdxjg\" (UniqueName: \"kubernetes.io/projected/feec8c4b-e98c-4fdd-bf70-235b413d7e90-kube-api-access-pdxjg\") on node \"crc\" DevicePath \"\"" Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.828992 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-x7vqg_eb8a54aa-2ce0-400c-a06f-cc3513b1df30/manager/0.log" Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.832585 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-x7vqg_eb8a54aa-2ce0-400c-a06f-cc3513b1df30/kube-rbac-proxy/0.log" Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.974993 4969 scope.go:117] "RemoveContainer" containerID="37b3249dedfa6f019e03735c72c5d26a9bd04c4863f8a1d657f1bb40bb317221" Oct 04 10:04:06 crc kubenswrapper[4969]: I1004 10:04:06.975130 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/crc-debug-q4tll" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.072196 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feec8c4b-e98c-4fdd-bf70-235b413d7e90" path="/var/lib/kubelet/pods/feec8c4b-e98c-4fdd-bf70-235b413d7e90/volumes" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.087383 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-d7885_d7dc7acf-3666-46fd-99cd-8447344cd10c/manager/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.095000 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-d7885_d7dc7acf-3666-46fd-99cd-8447344cd10c/kube-rbac-proxy/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.204173 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/util/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.410914 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/util/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.425430 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/pull/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.477302 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/pull/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.605352 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/util/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.636155 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/pull/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.660216 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d85ac18cfedb6a29c78319567a67294e0de443dd059c260ce07e4779becbcng_ed5c8e4a-dc27-4aa1-a533-9f918932ddff/extract/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.785211 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-m6llv_b7837ab8-145d-4f08-b407-5f0325119fb1/kube-rbac-proxy/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.847715 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-m6llv_b7837ab8-145d-4f08-b407-5f0325119fb1/manager/0.log" Oct 04 10:04:07 crc kubenswrapper[4969]: I1004 10:04:07.869293 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-lvh4x_0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f/kube-rbac-proxy/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.055535 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-lvh4x_0f62c1ca-3d74-4ee2-a95c-c378a84e7b9f/manager/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.084141 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-htxsn_fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea/kube-rbac-proxy/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.085030 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-htxsn_fc4ed288-8f9e-44a4-abda-81b7b4dfd8ea/manager/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.254145 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-m8tn8_18413b49-54b1-4ac9-8225-4aa748d0e4f3/kube-rbac-proxy/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.426825 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-m8tn8_18413b49-54b1-4ac9-8225-4aa748d0e4f3/manager/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.563330 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-zg6hn_89f37d97-0acc-4940-bd3a-733f4e6ed592/kube-rbac-proxy/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.649140 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-x8xxq_edfa72c9-187e-4d6c-8aab-66cb36073282/kube-rbac-proxy/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.733057 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-zg6hn_89f37d97-0acc-4940-bd3a-733f4e6ed592/manager/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.802251 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-x8xxq_edfa72c9-187e-4d6c-8aab-66cb36073282/manager/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.889227 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-t95db_5b816f82-7f2c-4537-aadd-53b6bded2e14/kube-rbac-proxy/0.log" Oct 04 10:04:08 crc kubenswrapper[4969]: I1004 10:04:08.991004 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-t95db_5b816f82-7f2c-4537-aadd-53b6bded2e14/manager/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.096688 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-lqq2s_d4376d0e-3428-49ec-993f-48b32d4c8863/kube-rbac-proxy/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.099804 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-lqq2s_d4376d0e-3428-49ec-993f-48b32d4c8863/manager/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.271951 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t_5603909a-7f2a-4c86-8446-9fae64c02482/kube-rbac-proxy/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.323892 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-pmf5t_5603909a-7f2a-4c86-8446-9fae64c02482/manager/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.394057 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-54zxr_0ce59178-f382-4436-be86-5bc4c5d4a9b0/kube-rbac-proxy/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.546531 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-54zxr_0ce59178-f382-4436-be86-5bc4c5d4a9b0/manager/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.578675 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-bg7zq_34735d37-aafa-4641-a951-f4d4d7fdab3b/kube-rbac-proxy/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.772119 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-bg7zq_34735d37-aafa-4641-a951-f4d4d7fdab3b/manager/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.809622 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-tf2wf_b457fc86-0c8f-4c98-9cbe-deddce25e44e/kube-rbac-proxy/0.log" Oct 04 10:04:09 crc kubenswrapper[4969]: I1004 10:04:09.815889 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-tf2wf_b457fc86-0c8f-4c98-9cbe-deddce25e44e/manager/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.022579 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g_24b8e975-2aeb-46ff-81e7-1b006bc37688/kube-rbac-proxy/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.046922 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cpn96g_24b8e975-2aeb-46ff-81e7-1b006bc37688/manager/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.192640 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-68dfdd7f5c-nl8g9_5523db3e-3b39-4bda-8885-898df0dcd5e7/kube-rbac-proxy/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.239273 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-76fbc76964-zng89_bccc7a17-2a0f-4fbb-80da-a1775673e40f/kube-rbac-proxy/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.473987 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-76fbc76964-zng89_bccc7a17-2a0f-4fbb-80da-a1775673e40f/operator/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.484917 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-d748d_03968d39-1ce4-4b22-875f-5b6f391a21c3/registry-server/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.624715 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-rq5fl_17114ec7-df18-4456-9e51-e93d74881841/kube-rbac-proxy/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.751892 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-rq5fl_17114ec7-df18-4456-9e51-e93d74881841/manager/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.799821 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-qxvcq_8e918774-159b-41dd-b320-5aaded1b8f52/kube-rbac-proxy/0.log" Oct 04 10:04:10 crc kubenswrapper[4969]: I1004 10:04:10.858017 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-qxvcq_8e918774-159b-41dd-b320-5aaded1b8f52/manager/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.009118 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-v8wxz_0b27d59e-dad0-402c-b7db-7d0dd87bd68e/operator/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.103342 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-jznzg_5c506bb2-6e70-425c-b4ed-be10eb472389/kube-rbac-proxy/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.235479 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-jznzg_5c506bb2-6e70-425c-b4ed-be10eb472389/manager/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.279798 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-kt6j5_d882bca8-0cc7-4053-8684-82707b083709/kube-rbac-proxy/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.335283 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-68dfdd7f5c-nl8g9_5523db3e-3b39-4bda-8885-898df0dcd5e7/manager/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.508029 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-ktq5d_8bf2e62a-8649-4b9e-8557-84441113bbdc/kube-rbac-proxy/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.550935 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-ktq5d_8bf2e62a-8649-4b9e-8557-84441113bbdc/manager/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.572082 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-kt6j5_d882bca8-0cc7-4053-8684-82707b083709/manager/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.663552 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-676569d79-24njs_a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9/kube-rbac-proxy/0.log" Oct 04 10:04:11 crc kubenswrapper[4969]: I1004 10:04:11.720361 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-676569d79-24njs_a65cbcc8-1908-47d5-b4bc-561f6dd8c9f9/manager/0.log" Oct 04 10:04:28 crc kubenswrapper[4969]: I1004 10:04:28.974597 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mtn99_d24a7bdf-1b12-49f8-b351-896f6edf2654/control-plane-machine-set-operator/0.log" Oct 04 10:04:29 crc kubenswrapper[4969]: I1004 10:04:29.114484 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-swbhw_5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8/kube-rbac-proxy/0.log" Oct 04 10:04:29 crc kubenswrapper[4969]: I1004 10:04:29.158219 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-swbhw_5bd319c3-7ee3-4897-b9cf-bc06c63ca3a8/machine-api-operator/0.log" Oct 04 10:04:41 crc kubenswrapper[4969]: I1004 10:04:41.993255 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-w94hv_21c43c6c-8aa0-4558-8e9e-ba34798b6c18/cert-manager-controller/0.log" Oct 04 10:04:42 crc kubenswrapper[4969]: I1004 10:04:42.093415 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-8s797_daf3ed47-3e82-4a2a-ae17-d56aa124499e/cert-manager-cainjector/0.log" Oct 04 10:04:42 crc kubenswrapper[4969]: I1004 10:04:42.168859 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-rdw9w_f37f93ad-8593-4bf2-bb4a-7d01e9ab273c/cert-manager-webhook/0.log" Oct 04 10:04:54 crc kubenswrapper[4969]: I1004 10:04:54.496949 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-mnl4l_7f4625c5-29dd-4cf4-8bb6-9bc88fe16a43/nmstate-console-plugin/0.log" Oct 04 10:04:54 crc kubenswrapper[4969]: I1004 10:04:54.705882 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5mjb6_eec45e0d-daf2-44e0-89d1-df0ccfe4fdd9/nmstate-handler/0.log" Oct 04 10:04:54 crc kubenswrapper[4969]: I1004 10:04:54.712315 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-t54hh_af2281a5-7ed5-4d7b-b1ab-f97948b7fd59/kube-rbac-proxy/0.log" Oct 04 10:04:54 crc kubenswrapper[4969]: I1004 10:04:54.758054 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-t54hh_af2281a5-7ed5-4d7b-b1ab-f97948b7fd59/nmstate-metrics/0.log" Oct 04 10:04:54 crc kubenswrapper[4969]: I1004 10:04:54.912616 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-p5kmf_005b44cf-b29b-4c9d-aabf-684a15eaba71/nmstate-operator/0.log" Oct 04 10:04:55 crc kubenswrapper[4969]: I1004 10:04:55.010394 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-rsbrd_dcb7ec91-3c05-474d-aead-c3014ca7e646/nmstate-webhook/0.log" Oct 04 10:05:10 crc kubenswrapper[4969]: I1004 10:05:10.405374 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cfvrx_b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b/kube-rbac-proxy/0.log" Oct 04 10:05:10 crc kubenswrapper[4969]: I1004 10:05:10.581924 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cfvrx_b2acffa8-6ecf-4c46-bcf8-d4a6852a7a5b/controller/0.log" Oct 04 10:05:10 crc kubenswrapper[4969]: I1004 10:05:10.642295 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-frr-files/0.log" Oct 04 10:05:10 crc kubenswrapper[4969]: I1004 10:05:10.795457 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-metrics/0.log" Oct 04 10:05:10 crc kubenswrapper[4969]: I1004 10:05:10.855033 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-frr-files/0.log" Oct 04 10:05:10 crc kubenswrapper[4969]: I1004 10:05:10.855589 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-reloader/0.log" Oct 04 10:05:10 crc kubenswrapper[4969]: I1004 10:05:10.855656 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-reloader/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.159852 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-metrics/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.209045 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-frr-files/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.240287 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-metrics/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.240460 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-reloader/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.449854 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/controller/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.456962 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-metrics/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.478577 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-frr-files/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.503202 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/cp-reloader/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.707585 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/kube-rbac-proxy/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.715306 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/frr-metrics/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.764867 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/kube-rbac-proxy-frr/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.917183 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/reloader/0.log" Oct 04 10:05:11 crc kubenswrapper[4969]: I1004 10:05:11.999908 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-pz8cc_897cdf44-f275-487d-85e6-9f47825f8b87/frr-k8s-webhook-server/0.log" Oct 04 10:05:12 crc kubenswrapper[4969]: I1004 10:05:12.122981 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-74c8fbc96b-hr67n_67b0a26f-7d5e-4988-83ba-8073480c2e2c/manager/0.log" Oct 04 10:05:12 crc kubenswrapper[4969]: I1004 10:05:12.351786 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6bd7b55ffd-6gshk_d0240b5e-b8bb-4690-a95a-f291b834c0d0/webhook-server/0.log" Oct 04 10:05:12 crc kubenswrapper[4969]: I1004 10:05:12.490433 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wtswn_7f4070ca-20a1-4c43-af82-890ef08c6fc3/kube-rbac-proxy/0.log" Oct 04 10:05:13 crc kubenswrapper[4969]: I1004 10:05:13.095470 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wtswn_7f4070ca-20a1-4c43-af82-890ef08c6fc3/speaker/0.log" Oct 04 10:05:13 crc kubenswrapper[4969]: I1004 10:05:13.558879 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wdwv9_6d4150d2-810e-4ceb-86a0-59155ecb1b0d/frr/0.log" Oct 04 10:05:26 crc kubenswrapper[4969]: I1004 10:05:26.644173 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/util/0.log" Oct 04 10:05:26 crc kubenswrapper[4969]: I1004 10:05:26.762342 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/pull/0.log" Oct 04 10:05:26 crc kubenswrapper[4969]: I1004 10:05:26.788119 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/util/0.log" Oct 04 10:05:26 crc kubenswrapper[4969]: I1004 10:05:26.814342 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/pull/0.log" Oct 04 10:05:26 crc kubenswrapper[4969]: I1004 10:05:26.981645 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/extract/0.log" Oct 04 10:05:26 crc kubenswrapper[4969]: I1004 10:05:26.997669 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/util/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.051930 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2jpcb5_71587e21-d80e-404c-8ccf-aff572933817/pull/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.168702 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/util/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.308193 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/util/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.316702 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/pull/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.332849 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/pull/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.480559 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/util/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.504722 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/extract/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.512017 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d5mmvw_82b1344f-981d-4caf-82c2-c8035d9fb68b/pull/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.655946 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-utilities/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.818796 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-utilities/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.850276 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-content/0.log" Oct 04 10:05:27 crc kubenswrapper[4969]: I1004 10:05:27.860853 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-content/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.008773 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-content/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.010992 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/extract-utilities/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.221363 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-utilities/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.464114 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-content/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.488737 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-content/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.493459 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-utilities/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.705127 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-utilities/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.737750 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/extract-content/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.785544 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwlvl_d378dd74-7b3b-4e78-9815-f75e40f48d1f/registry-server/0.log" Oct 04 10:05:28 crc kubenswrapper[4969]: I1004 10:05:28.985549 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/util/0.log" Oct 04 10:05:29 crc kubenswrapper[4969]: I1004 10:05:29.188324 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/util/0.log" Oct 04 10:05:29 crc kubenswrapper[4969]: I1004 10:05:29.256822 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/pull/0.log" Oct 04 10:05:29 crc kubenswrapper[4969]: I1004 10:05:29.388213 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/pull/0.log" Oct 04 10:05:29 crc kubenswrapper[4969]: I1004 10:05:29.575186 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5drz6_d8276ca7-275d-4cff-8895-f868dab88c73/registry-server/0.log" Oct 04 10:05:29 crc kubenswrapper[4969]: I1004 10:05:29.618795 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/util/0.log" Oct 04 10:05:29 crc kubenswrapper[4969]: I1004 10:05:29.642715 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/pull/0.log" Oct 04 10:05:29 crc kubenswrapper[4969]: I1004 10:05:29.648330 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdsdhp_12c98149-2fa0-4b54-9940-92d1a818024c/extract/0.log" Oct 04 10:05:29 crc kubenswrapper[4969]: I1004 10:05:29.847094 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-utilities/0.log" Oct 04 10:05:29 crc kubenswrapper[4969]: I1004 10:05:29.854190 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5qth5_0cf38a30-f714-4a15-8be5-30118e8984c7/marketplace-operator/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.003438 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-utilities/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.021686 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-content/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.025211 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-content/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.226403 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-content/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.226847 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/extract-utilities/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.303386 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-utilities/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.439169 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-content/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.454807 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xrjgj_34db4256-784d-48a4-ae67-47e33b1c0e9f/registry-server/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.461260 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-content/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.509130 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-utilities/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.664829 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-content/0.log" Oct 04 10:05:30 crc kubenswrapper[4969]: I1004 10:05:30.678309 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/extract-utilities/0.log" Oct 04 10:05:31 crc kubenswrapper[4969]: I1004 10:05:31.359557 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p7tnv_3a59cef3-965a-4244-89af-0eaa80eef618/registry-server/0.log" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.158356 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5zst8"] Oct 04 10:05:36 crc kubenswrapper[4969]: E1004 10:05:36.159317 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feec8c4b-e98c-4fdd-bf70-235b413d7e90" containerName="container-00" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.159339 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="feec8c4b-e98c-4fdd-bf70-235b413d7e90" containerName="container-00" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.159717 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="feec8c4b-e98c-4fdd-bf70-235b413d7e90" containerName="container-00" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.161624 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.166249 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5zst8"] Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.288813 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-utilities\") pod \"community-operators-5zst8\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.289136 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-catalog-content\") pod \"community-operators-5zst8\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.289172 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnhsl\" (UniqueName: \"kubernetes.io/projected/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-kube-api-access-gnhsl\") pod \"community-operators-5zst8\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.390959 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-utilities\") pod \"community-operators-5zst8\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.391264 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-catalog-content\") pod \"community-operators-5zst8\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.391413 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnhsl\" (UniqueName: \"kubernetes.io/projected/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-kube-api-access-gnhsl\") pod \"community-operators-5zst8\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.391556 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-utilities\") pod \"community-operators-5zst8\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.391785 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-catalog-content\") pod \"community-operators-5zst8\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.414458 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnhsl\" (UniqueName: \"kubernetes.io/projected/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-kube-api-access-gnhsl\") pod \"community-operators-5zst8\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.502811 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:36 crc kubenswrapper[4969]: I1004 10:05:36.991391 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5zst8"] Oct 04 10:05:37 crc kubenswrapper[4969]: I1004 10:05:37.945509 4969 generic.go:334] "Generic (PLEG): container finished" podID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerID="e636ec2276307e257aaa5b4c3ace4cfbcbae3bc1dfb0b5e40b9f4fbe85317375" exitCode=0 Oct 04 10:05:37 crc kubenswrapper[4969]: I1004 10:05:37.945724 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zst8" event={"ID":"aa6acb8d-4cc7-42dc-8f41-0c74700985dd","Type":"ContainerDied","Data":"e636ec2276307e257aaa5b4c3ace4cfbcbae3bc1dfb0b5e40b9f4fbe85317375"} Oct 04 10:05:37 crc kubenswrapper[4969]: I1004 10:05:37.945877 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zst8" event={"ID":"aa6acb8d-4cc7-42dc-8f41-0c74700985dd","Type":"ContainerStarted","Data":"fbf9cfb5c64e82e69d682c215a17301967e390f9512555a49ee61624255ecf0d"} Oct 04 10:05:37 crc kubenswrapper[4969]: I1004 10:05:37.949285 4969 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 10:05:38 crc kubenswrapper[4969]: I1004 10:05:38.957738 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zst8" event={"ID":"aa6acb8d-4cc7-42dc-8f41-0c74700985dd","Type":"ContainerStarted","Data":"d82dcea20f0dc859dca90baf34e3d475362542d9685d0d554cab4431e92363af"} Oct 04 10:05:39 crc kubenswrapper[4969]: I1004 10:05:39.973653 4969 generic.go:334] "Generic (PLEG): container finished" podID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerID="d82dcea20f0dc859dca90baf34e3d475362542d9685d0d554cab4431e92363af" exitCode=0 Oct 04 10:05:39 crc kubenswrapper[4969]: I1004 10:05:39.973776 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zst8" event={"ID":"aa6acb8d-4cc7-42dc-8f41-0c74700985dd","Type":"ContainerDied","Data":"d82dcea20f0dc859dca90baf34e3d475362542d9685d0d554cab4431e92363af"} Oct 04 10:05:40 crc kubenswrapper[4969]: I1004 10:05:40.985707 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zst8" event={"ID":"aa6acb8d-4cc7-42dc-8f41-0c74700985dd","Type":"ContainerStarted","Data":"2c697e23748b44b2ac2cbb0848d9ab20d961bfc368f59efd7733d19ee6105728"} Oct 04 10:05:41 crc kubenswrapper[4969]: I1004 10:05:41.004086 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5zst8" podStartSLOduration=2.5245937769999998 podStartE2EDuration="5.004066011s" podCreationTimestamp="2025-10-04 10:05:36 +0000 UTC" firstStartedPulling="2025-10-04 10:05:37.948935498 +0000 UTC m=+6565.703204332" lastFinishedPulling="2025-10-04 10:05:40.428407752 +0000 UTC m=+6568.182676566" observedRunningTime="2025-10-04 10:05:40.999191751 +0000 UTC m=+6568.753460615" watchObservedRunningTime="2025-10-04 10:05:41.004066011 +0000 UTC m=+6568.758334825" Oct 04 10:05:44 crc kubenswrapper[4969]: I1004 10:05:44.149802 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-2xmcb_91044ecd-9c81-4052-9e16-9224f04f079e/prometheus-operator/0.log" Oct 04 10:05:44 crc kubenswrapper[4969]: I1004 10:05:44.307565 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7746cfc87b-tsxj6_fca6110d-f0fb-4b9e-8c48-749c188c96a3/prometheus-operator-admission-webhook/0.log" Oct 04 10:05:44 crc kubenswrapper[4969]: I1004 10:05:44.366480 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7746cfc87b-w75mn_c7ba6dc9-cfcc-4e19-8502-b5ef4c981b1e/prometheus-operator-admission-webhook/0.log" Oct 04 10:05:44 crc kubenswrapper[4969]: I1004 10:05:44.509834 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-2c6cm_46090d90-0cf0-47f4-a19f-ad5aa65df521/perses-operator/0.log" Oct 04 10:05:44 crc kubenswrapper[4969]: I1004 10:05:44.512272 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-vl7q6_15b3f66e-d095-42c3-94e2-b9ee1071d638/operator/0.log" Oct 04 10:05:46 crc kubenswrapper[4969]: I1004 10:05:46.503305 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:46 crc kubenswrapper[4969]: I1004 10:05:46.503601 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:46 crc kubenswrapper[4969]: I1004 10:05:46.573372 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:47 crc kubenswrapper[4969]: I1004 10:05:47.123126 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:47 crc kubenswrapper[4969]: I1004 10:05:47.176134 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5zst8"] Oct 04 10:05:49 crc kubenswrapper[4969]: I1004 10:05:49.059661 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5zst8" podUID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerName="registry-server" containerID="cri-o://2c697e23748b44b2ac2cbb0848d9ab20d961bfc368f59efd7733d19ee6105728" gracePeriod=2 Oct 04 10:05:49 crc kubenswrapper[4969]: E1004 10:05:49.605283 4969 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa6acb8d_4cc7_42dc_8f41_0c74700985dd.slice/crio-conmon-2c697e23748b44b2ac2cbb0848d9ab20d961bfc368f59efd7733d19ee6105728.scope\": RecentStats: unable to find data in memory cache]" Oct 04 10:05:49 crc kubenswrapper[4969]: I1004 10:05:49.666868 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:05:49 crc kubenswrapper[4969]: I1004 10:05:49.666947 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.077596 4969 generic.go:334] "Generic (PLEG): container finished" podID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerID="2c697e23748b44b2ac2cbb0848d9ab20d961bfc368f59efd7733d19ee6105728" exitCode=0 Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.077657 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zst8" event={"ID":"aa6acb8d-4cc7-42dc-8f41-0c74700985dd","Type":"ContainerDied","Data":"2c697e23748b44b2ac2cbb0848d9ab20d961bfc368f59efd7733d19ee6105728"} Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.077930 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zst8" event={"ID":"aa6acb8d-4cc7-42dc-8f41-0c74700985dd","Type":"ContainerDied","Data":"fbf9cfb5c64e82e69d682c215a17301967e390f9512555a49ee61624255ecf0d"} Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.077946 4969 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbf9cfb5c64e82e69d682c215a17301967e390f9512555a49ee61624255ecf0d" Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.140049 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.320551 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnhsl\" (UniqueName: \"kubernetes.io/projected/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-kube-api-access-gnhsl\") pod \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.320886 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-utilities\") pod \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.320942 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-catalog-content\") pod \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\" (UID: \"aa6acb8d-4cc7-42dc-8f41-0c74700985dd\") " Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.322102 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-utilities" (OuterVolumeSpecName: "utilities") pod "aa6acb8d-4cc7-42dc-8f41-0c74700985dd" (UID: "aa6acb8d-4cc7-42dc-8f41-0c74700985dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.326397 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-kube-api-access-gnhsl" (OuterVolumeSpecName: "kube-api-access-gnhsl") pod "aa6acb8d-4cc7-42dc-8f41-0c74700985dd" (UID: "aa6acb8d-4cc7-42dc-8f41-0c74700985dd"). InnerVolumeSpecName "kube-api-access-gnhsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.365853 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa6acb8d-4cc7-42dc-8f41-0c74700985dd" (UID: "aa6acb8d-4cc7-42dc-8f41-0c74700985dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.422710 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.422742 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnhsl\" (UniqueName: \"kubernetes.io/projected/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-kube-api-access-gnhsl\") on node \"crc\" DevicePath \"\"" Oct 04 10:05:50 crc kubenswrapper[4969]: I1004 10:05:50.422752 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6acb8d-4cc7-42dc-8f41-0c74700985dd-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:05:51 crc kubenswrapper[4969]: I1004 10:05:51.084484 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zst8" Oct 04 10:05:51 crc kubenswrapper[4969]: I1004 10:05:51.106107 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5zst8"] Oct 04 10:05:51 crc kubenswrapper[4969]: I1004 10:05:51.119171 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5zst8"] Oct 04 10:05:53 crc kubenswrapper[4969]: I1004 10:05:53.068631 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" path="/var/lib/kubelet/pods/aa6acb8d-4cc7-42dc-8f41-0c74700985dd/volumes" Oct 04 10:05:55 crc kubenswrapper[4969]: E1004 10:05:55.640724 4969 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.195:56664->38.102.83.195:43269: write tcp 38.102.83.195:56664->38.102.83.195:43269: write: broken pipe Oct 04 10:06:19 crc kubenswrapper[4969]: I1004 10:06:19.666728 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:06:19 crc kubenswrapper[4969]: I1004 10:06:19.667373 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:06:49 crc kubenswrapper[4969]: I1004 10:06:49.667071 4969 patch_prober.go:28] interesting pod/machine-config-daemon-bc2w6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:06:49 crc kubenswrapper[4969]: I1004 10:06:49.667668 4969 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:06:49 crc kubenswrapper[4969]: I1004 10:06:49.667720 4969 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" Oct 04 10:06:49 crc kubenswrapper[4969]: I1004 10:06:49.670303 4969 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b"} pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:06:49 crc kubenswrapper[4969]: I1004 10:06:49.670474 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerName="machine-config-daemon" containerID="cri-o://779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" gracePeriod=600 Oct 04 10:06:49 crc kubenswrapper[4969]: E1004 10:06:49.825411 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:06:50 crc kubenswrapper[4969]: I1004 10:06:50.829877 4969 generic.go:334] "Generic (PLEG): container finished" podID="69d1f843-03d6-403f-8ab3-796e2c97af4f" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" exitCode=0 Oct 04 10:06:50 crc kubenswrapper[4969]: I1004 10:06:50.829984 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" event={"ID":"69d1f843-03d6-403f-8ab3-796e2c97af4f","Type":"ContainerDied","Data":"779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b"} Oct 04 10:06:50 crc kubenswrapper[4969]: I1004 10:06:50.830193 4969 scope.go:117] "RemoveContainer" containerID="8fdf0e4cb48a4d351d146bddccf73a2a4695a141ecf2ec805971c3bfed433381" Oct 04 10:06:50 crc kubenswrapper[4969]: I1004 10:06:50.830844 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:06:50 crc kubenswrapper[4969]: E1004 10:06:50.831213 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:07:02 crc kubenswrapper[4969]: I1004 10:07:02.056605 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:07:02 crc kubenswrapper[4969]: E1004 10:07:02.057784 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.012963 4969 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xwns5"] Oct 04 10:07:14 crc kubenswrapper[4969]: E1004 10:07:14.013978 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerName="registry-server" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.013993 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerName="registry-server" Oct 04 10:07:14 crc kubenswrapper[4969]: E1004 10:07:14.014007 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerName="extract-utilities" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.014016 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerName="extract-utilities" Oct 04 10:07:14 crc kubenswrapper[4969]: E1004 10:07:14.014044 4969 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerName="extract-content" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.014054 4969 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerName="extract-content" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.014303 4969 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa6acb8d-4cc7-42dc-8f41-0c74700985dd" containerName="registry-server" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.016690 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.028212 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwns5"] Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.096337 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-utilities\") pod \"redhat-operators-xwns5\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.096406 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp65d\" (UniqueName: \"kubernetes.io/projected/83a39415-5ff1-4b1d-b37b-c31ada5d982c-kube-api-access-pp65d\") pod \"redhat-operators-xwns5\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.096645 4969 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-catalog-content\") pod \"redhat-operators-xwns5\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.198312 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-catalog-content\") pod \"redhat-operators-xwns5\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.198385 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-utilities\") pod \"redhat-operators-xwns5\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.198473 4969 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp65d\" (UniqueName: \"kubernetes.io/projected/83a39415-5ff1-4b1d-b37b-c31ada5d982c-kube-api-access-pp65d\") pod \"redhat-operators-xwns5\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.199178 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-catalog-content\") pod \"redhat-operators-xwns5\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.199242 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-utilities\") pod \"redhat-operators-xwns5\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.222092 4969 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp65d\" (UniqueName: \"kubernetes.io/projected/83a39415-5ff1-4b1d-b37b-c31ada5d982c-kube-api-access-pp65d\") pod \"redhat-operators-xwns5\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.348253 4969 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:14 crc kubenswrapper[4969]: I1004 10:07:14.844932 4969 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwns5"] Oct 04 10:07:15 crc kubenswrapper[4969]: I1004 10:07:15.055221 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:07:15 crc kubenswrapper[4969]: E1004 10:07:15.055486 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:07:15 crc kubenswrapper[4969]: I1004 10:07:15.170529 4969 generic.go:334] "Generic (PLEG): container finished" podID="83a39415-5ff1-4b1d-b37b-c31ada5d982c" containerID="3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee" exitCode=0 Oct 04 10:07:15 crc kubenswrapper[4969]: I1004 10:07:15.170742 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwns5" event={"ID":"83a39415-5ff1-4b1d-b37b-c31ada5d982c","Type":"ContainerDied","Data":"3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee"} Oct 04 10:07:15 crc kubenswrapper[4969]: I1004 10:07:15.170767 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwns5" event={"ID":"83a39415-5ff1-4b1d-b37b-c31ada5d982c","Type":"ContainerStarted","Data":"72e4114f3632375e2a948c8fb595ca0d2f5dd0c5cdcc4d6406be27d6b547e387"} Oct 04 10:07:17 crc kubenswrapper[4969]: I1004 10:07:17.196002 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwns5" event={"ID":"83a39415-5ff1-4b1d-b37b-c31ada5d982c","Type":"ContainerStarted","Data":"83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0"} Oct 04 10:07:18 crc kubenswrapper[4969]: I1004 10:07:18.215703 4969 generic.go:334] "Generic (PLEG): container finished" podID="83a39415-5ff1-4b1d-b37b-c31ada5d982c" containerID="83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0" exitCode=0 Oct 04 10:07:18 crc kubenswrapper[4969]: I1004 10:07:18.215824 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwns5" event={"ID":"83a39415-5ff1-4b1d-b37b-c31ada5d982c","Type":"ContainerDied","Data":"83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0"} Oct 04 10:07:19 crc kubenswrapper[4969]: I1004 10:07:19.226735 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwns5" event={"ID":"83a39415-5ff1-4b1d-b37b-c31ada5d982c","Type":"ContainerStarted","Data":"68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67"} Oct 04 10:07:24 crc kubenswrapper[4969]: I1004 10:07:24.353048 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:24 crc kubenswrapper[4969]: I1004 10:07:24.353682 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:25 crc kubenswrapper[4969]: I1004 10:07:25.442323 4969 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xwns5" podUID="83a39415-5ff1-4b1d-b37b-c31ada5d982c" containerName="registry-server" probeResult="failure" output=< Oct 04 10:07:25 crc kubenswrapper[4969]: timeout: failed to connect service ":50051" within 1s Oct 04 10:07:25 crc kubenswrapper[4969]: > Oct 04 10:07:27 crc kubenswrapper[4969]: I1004 10:07:27.056213 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:07:27 crc kubenswrapper[4969]: E1004 10:07:27.056973 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:07:34 crc kubenswrapper[4969]: I1004 10:07:34.432560 4969 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:34 crc kubenswrapper[4969]: I1004 10:07:34.460147 4969 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xwns5" podStartSLOduration=17.843179663 podStartE2EDuration="21.460129636s" podCreationTimestamp="2025-10-04 10:07:13 +0000 UTC" firstStartedPulling="2025-10-04 10:07:15.172362949 +0000 UTC m=+6662.926631763" lastFinishedPulling="2025-10-04 10:07:18.789312912 +0000 UTC m=+6666.543581736" observedRunningTime="2025-10-04 10:07:19.259771277 +0000 UTC m=+6667.014040091" watchObservedRunningTime="2025-10-04 10:07:34.460129636 +0000 UTC m=+6682.214398460" Oct 04 10:07:34 crc kubenswrapper[4969]: I1004 10:07:34.517806 4969 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:34 crc kubenswrapper[4969]: I1004 10:07:34.677130 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwns5"] Oct 04 10:07:36 crc kubenswrapper[4969]: I1004 10:07:36.467323 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xwns5" podUID="83a39415-5ff1-4b1d-b37b-c31ada5d982c" containerName="registry-server" containerID="cri-o://68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67" gracePeriod=2 Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.028213 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.140325 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-catalog-content\") pod \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.140528 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-utilities\") pod \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.140645 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp65d\" (UniqueName: \"kubernetes.io/projected/83a39415-5ff1-4b1d-b37b-c31ada5d982c-kube-api-access-pp65d\") pod \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.142851 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-utilities" (OuterVolumeSpecName: "utilities") pod "83a39415-5ff1-4b1d-b37b-c31ada5d982c" (UID: "83a39415-5ff1-4b1d-b37b-c31ada5d982c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.147767 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a39415-5ff1-4b1d-b37b-c31ada5d982c-kube-api-access-pp65d" (OuterVolumeSpecName: "kube-api-access-pp65d") pod "83a39415-5ff1-4b1d-b37b-c31ada5d982c" (UID: "83a39415-5ff1-4b1d-b37b-c31ada5d982c"). InnerVolumeSpecName "kube-api-access-pp65d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.242718 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83a39415-5ff1-4b1d-b37b-c31ada5d982c" (UID: "83a39415-5ff1-4b1d-b37b-c31ada5d982c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.243237 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-catalog-content\") pod \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\" (UID: \"83a39415-5ff1-4b1d-b37b-c31ada5d982c\") " Oct 04 10:07:37 crc kubenswrapper[4969]: W1004 10:07:37.243350 4969 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/83a39415-5ff1-4b1d-b37b-c31ada5d982c/volumes/kubernetes.io~empty-dir/catalog-content Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.243366 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83a39415-5ff1-4b1d-b37b-c31ada5d982c" (UID: "83a39415-5ff1-4b1d-b37b-c31ada5d982c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.244106 4969 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.244131 4969 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a39415-5ff1-4b1d-b37b-c31ada5d982c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.244148 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp65d\" (UniqueName: \"kubernetes.io/projected/83a39415-5ff1-4b1d-b37b-c31ada5d982c-kube-api-access-pp65d\") on node \"crc\" DevicePath \"\"" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.482480 4969 generic.go:334] "Generic (PLEG): container finished" podID="83a39415-5ff1-4b1d-b37b-c31ada5d982c" containerID="68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67" exitCode=0 Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.482525 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwns5" event={"ID":"83a39415-5ff1-4b1d-b37b-c31ada5d982c","Type":"ContainerDied","Data":"68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67"} Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.482558 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwns5" event={"ID":"83a39415-5ff1-4b1d-b37b-c31ada5d982c","Type":"ContainerDied","Data":"72e4114f3632375e2a948c8fb595ca0d2f5dd0c5cdcc4d6406be27d6b547e387"} Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.482576 4969 scope.go:117] "RemoveContainer" containerID="68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.482592 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwns5" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.504308 4969 scope.go:117] "RemoveContainer" containerID="83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.535739 4969 scope.go:117] "RemoveContainer" containerID="3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.541307 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwns5"] Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.550073 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xwns5"] Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.600619 4969 scope.go:117] "RemoveContainer" containerID="68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67" Oct 04 10:07:37 crc kubenswrapper[4969]: E1004 10:07:37.600932 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67\": container with ID starting with 68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67 not found: ID does not exist" containerID="68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.600963 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67"} err="failed to get container status \"68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67\": rpc error: code = NotFound desc = could not find container \"68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67\": container with ID starting with 68f07eedbfdd7ce064319d97147cceb257bd338238f0da44b826ed30fb482e67 not found: ID does not exist" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.600984 4969 scope.go:117] "RemoveContainer" containerID="83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0" Oct 04 10:07:37 crc kubenswrapper[4969]: E1004 10:07:37.601189 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0\": container with ID starting with 83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0 not found: ID does not exist" containerID="83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.601210 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0"} err="failed to get container status \"83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0\": rpc error: code = NotFound desc = could not find container \"83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0\": container with ID starting with 83b2ac0a55ceabee3905d3fc707f4b3a67f0c982d9ef25717b48ba7646f229b0 not found: ID does not exist" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.601224 4969 scope.go:117] "RemoveContainer" containerID="3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee" Oct 04 10:07:37 crc kubenswrapper[4969]: E1004 10:07:37.601485 4969 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee\": container with ID starting with 3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee not found: ID does not exist" containerID="3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee" Oct 04 10:07:37 crc kubenswrapper[4969]: I1004 10:07:37.601509 4969 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee"} err="failed to get container status \"3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee\": rpc error: code = NotFound desc = could not find container \"3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee\": container with ID starting with 3bd1ace31ed14301f393548de6f97217a18456ba681326a95ef663adf0f661ee not found: ID does not exist" Oct 04 10:07:38 crc kubenswrapper[4969]: I1004 10:07:38.055406 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:07:38 crc kubenswrapper[4969]: E1004 10:07:38.055904 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:07:39 crc kubenswrapper[4969]: I1004 10:07:39.074639 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a39415-5ff1-4b1d-b37b-c31ada5d982c" path="/var/lib/kubelet/pods/83a39415-5ff1-4b1d-b37b-c31ada5d982c/volumes" Oct 04 10:07:49 crc kubenswrapper[4969]: I1004 10:07:49.055930 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:07:49 crc kubenswrapper[4969]: E1004 10:07:49.057088 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:07:53 crc kubenswrapper[4969]: I1004 10:07:53.695089 4969 generic.go:334] "Generic (PLEG): container finished" podID="cbdba948-d8b5-4828-b27d-d9a93a49717a" containerID="f0b0ffae271acfc0b781ee4ace2439b173ab1ff54b9a2f1eb663a69b67d1dd05" exitCode=0 Oct 04 10:07:53 crc kubenswrapper[4969]: I1004 10:07:53.695249 4969 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" event={"ID":"cbdba948-d8b5-4828-b27d-d9a93a49717a","Type":"ContainerDied","Data":"f0b0ffae271acfc0b781ee4ace2439b173ab1ff54b9a2f1eb663a69b67d1dd05"} Oct 04 10:07:53 crc kubenswrapper[4969]: I1004 10:07:53.696584 4969 scope.go:117] "RemoveContainer" containerID="f0b0ffae271acfc0b781ee4ace2439b173ab1ff54b9a2f1eb663a69b67d1dd05" Oct 04 10:07:53 crc kubenswrapper[4969]: I1004 10:07:53.828600 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6b4sb_must-gather-jbgx9_cbdba948-d8b5-4828-b27d-d9a93a49717a/gather/0.log" Oct 04 10:08:00 crc kubenswrapper[4969]: I1004 10:08:00.055856 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:08:00 crc kubenswrapper[4969]: E1004 10:08:00.056997 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:08:07 crc kubenswrapper[4969]: I1004 10:08:07.538374 4969 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6b4sb/must-gather-jbgx9"] Oct 04 10:08:07 crc kubenswrapper[4969]: I1004 10:08:07.538961 4969 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" podUID="cbdba948-d8b5-4828-b27d-d9a93a49717a" containerName="copy" containerID="cri-o://338011150e91a2246896914e7107c5f2e91f5523aa3a99007693cf04675f48d4" gracePeriod=2 Oct 04 10:08:07 crc kubenswrapper[4969]: I1004 10:08:07.551249 4969 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6b4sb/must-gather-jbgx9"] Oct 04 10:08:07 crc kubenswrapper[4969]: I1004 10:08:07.875280 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6b4sb_must-gather-jbgx9_cbdba948-d8b5-4828-b27d-d9a93a49717a/copy/0.log" Oct 04 10:08:07 crc kubenswrapper[4969]: I1004 10:08:07.877361 4969 generic.go:334] "Generic (PLEG): container finished" podID="cbdba948-d8b5-4828-b27d-d9a93a49717a" containerID="338011150e91a2246896914e7107c5f2e91f5523aa3a99007693cf04675f48d4" exitCode=143 Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.024210 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6b4sb_must-gather-jbgx9_cbdba948-d8b5-4828-b27d-d9a93a49717a/copy/0.log" Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.024894 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.214246 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cbdba948-d8b5-4828-b27d-d9a93a49717a-must-gather-output\") pod \"cbdba948-d8b5-4828-b27d-d9a93a49717a\" (UID: \"cbdba948-d8b5-4828-b27d-d9a93a49717a\") " Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.214501 4969 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clfd8\" (UniqueName: \"kubernetes.io/projected/cbdba948-d8b5-4828-b27d-d9a93a49717a-kube-api-access-clfd8\") pod \"cbdba948-d8b5-4828-b27d-d9a93a49717a\" (UID: \"cbdba948-d8b5-4828-b27d-d9a93a49717a\") " Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.221662 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbdba948-d8b5-4828-b27d-d9a93a49717a-kube-api-access-clfd8" (OuterVolumeSpecName: "kube-api-access-clfd8") pod "cbdba948-d8b5-4828-b27d-d9a93a49717a" (UID: "cbdba948-d8b5-4828-b27d-d9a93a49717a"). InnerVolumeSpecName "kube-api-access-clfd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.317588 4969 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clfd8\" (UniqueName: \"kubernetes.io/projected/cbdba948-d8b5-4828-b27d-d9a93a49717a-kube-api-access-clfd8\") on node \"crc\" DevicePath \"\"" Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.425969 4969 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbdba948-d8b5-4828-b27d-d9a93a49717a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cbdba948-d8b5-4828-b27d-d9a93a49717a" (UID: "cbdba948-d8b5-4828-b27d-d9a93a49717a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.522901 4969 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cbdba948-d8b5-4828-b27d-d9a93a49717a-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.895723 4969 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6b4sb_must-gather-jbgx9_cbdba948-d8b5-4828-b27d-d9a93a49717a/copy/0.log" Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.900016 4969 scope.go:117] "RemoveContainer" containerID="338011150e91a2246896914e7107c5f2e91f5523aa3a99007693cf04675f48d4" Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.900050 4969 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6b4sb/must-gather-jbgx9" Oct 04 10:08:08 crc kubenswrapper[4969]: I1004 10:08:08.928301 4969 scope.go:117] "RemoveContainer" containerID="f0b0ffae271acfc0b781ee4ace2439b173ab1ff54b9a2f1eb663a69b67d1dd05" Oct 04 10:08:09 crc kubenswrapper[4969]: I1004 10:08:09.088141 4969 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbdba948-d8b5-4828-b27d-d9a93a49717a" path="/var/lib/kubelet/pods/cbdba948-d8b5-4828-b27d-d9a93a49717a/volumes" Oct 04 10:08:13 crc kubenswrapper[4969]: I1004 10:08:13.091936 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:08:13 crc kubenswrapper[4969]: E1004 10:08:13.094314 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:08:27 crc kubenswrapper[4969]: I1004 10:08:27.055217 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:08:27 crc kubenswrapper[4969]: E1004 10:08:27.057014 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:08:35 crc kubenswrapper[4969]: I1004 10:08:35.592675 4969 scope.go:117] "RemoveContainer" containerID="e493485269fc5f984c92b2cb126e270bdf3bad54200f121d9b0bb3b5ee8007ea" Oct 04 10:08:40 crc kubenswrapper[4969]: I1004 10:08:40.055611 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:08:40 crc kubenswrapper[4969]: E1004 10:08:40.056267 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:08:53 crc kubenswrapper[4969]: I1004 10:08:53.072882 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:08:53 crc kubenswrapper[4969]: E1004 10:08:53.074012 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:09:04 crc kubenswrapper[4969]: I1004 10:09:04.055708 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:09:04 crc kubenswrapper[4969]: E1004 10:09:04.056862 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:09:18 crc kubenswrapper[4969]: I1004 10:09:18.055347 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:09:18 crc kubenswrapper[4969]: E1004 10:09:18.056627 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:09:29 crc kubenswrapper[4969]: I1004 10:09:29.056393 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:09:29 crc kubenswrapper[4969]: E1004 10:09:29.057649 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" Oct 04 10:09:42 crc kubenswrapper[4969]: I1004 10:09:42.056934 4969 scope.go:117] "RemoveContainer" containerID="779f820d7a5005ecc58b9f6c187c6f0e1243b434cf2c972f7040dc0bb4cb8e3b" Oct 04 10:09:42 crc kubenswrapper[4969]: E1004 10:09:42.058090 4969 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bc2w6_openshift-machine-config-operator(69d1f843-03d6-403f-8ab3-796e2c97af4f)\"" pod="openshift-machine-config-operator/machine-config-daemon-bc2w6" podUID="69d1f843-03d6-403f-8ab3-796e2c97af4f" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070171163024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070171164017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070153351016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070153352015456 5ustar corecore